var/home/core/zuul-output/0000755000175000017500000000000015067444602014535 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015067457331015504 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005146775615067457322017735 0ustar rootrootOct 02 09:48:47 crc systemd[1]: Starting Kubernetes Kubelet... Oct 02 09:48:47 crc restorecon[4744]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:47 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:48:48 crc restorecon[4744]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:48:48 crc restorecon[4744]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 02 09:48:48 crc kubenswrapper[4934]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 09:48:48 crc kubenswrapper[4934]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 02 09:48:48 crc kubenswrapper[4934]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 09:48:48 crc kubenswrapper[4934]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 09:48:48 crc kubenswrapper[4934]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 02 09:48:48 crc kubenswrapper[4934]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.683062 4934 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688468 4934 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688484 4934 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688489 4934 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688493 4934 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688498 4934 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688501 4934 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688505 4934 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688508 4934 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688511 4934 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688515 4934 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688519 4934 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688522 4934 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688525 4934 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688529 4934 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688532 4934 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688536 4934 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688539 4934 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688543 4934 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688546 4934 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688550 4934 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688555 4934 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688558 4934 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688562 4934 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688565 4934 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688570 4934 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688588 4934 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688592 4934 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688596 4934 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688599 4934 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688603 4934 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688607 4934 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688611 4934 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688614 4934 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688617 4934 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688621 4934 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688625 4934 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688628 4934 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688632 4934 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688637 4934 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688640 4934 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688644 4934 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688647 4934 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688651 4934 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688655 4934 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688659 4934 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688663 4934 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688667 4934 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688670 4934 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688673 4934 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688677 4934 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688680 4934 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688684 4934 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688687 4934 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688692 4934 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688696 4934 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688700 4934 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688703 4934 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688707 4934 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688711 4934 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688714 4934 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688719 4934 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688723 4934 feature_gate.go:330] unrecognized feature gate: Example Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688728 4934 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688733 4934 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688736 4934 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688740 4934 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688744 4934 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688748 4934 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688751 4934 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688755 4934 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.688758 4934 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.688855 4934 flags.go:64] FLAG: --address="0.0.0.0" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.688877 4934 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.688886 4934 flags.go:64] FLAG: --anonymous-auth="true" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.688892 4934 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.688899 4934 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.688903 4934 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.688909 4934 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.688915 4934 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.688919 4934 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.688924 4934 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.688928 4934 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.688932 4934 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.688936 4934 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.688941 4934 flags.go:64] FLAG: --cgroup-root="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.688945 4934 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.688949 4934 flags.go:64] FLAG: --client-ca-file="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.688953 4934 flags.go:64] FLAG: --cloud-config="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.688957 4934 flags.go:64] FLAG: --cloud-provider="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.688961 4934 flags.go:64] FLAG: --cluster-dns="[]" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.688967 4934 flags.go:64] FLAG: --cluster-domain="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.688971 4934 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.688975 4934 flags.go:64] FLAG: --config-dir="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.688979 4934 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.688984 4934 flags.go:64] FLAG: --container-log-max-files="5" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.688989 4934 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.688993 4934 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.688998 4934 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689002 4934 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689006 4934 flags.go:64] FLAG: --contention-profiling="false" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689011 4934 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689015 4934 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689019 4934 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689023 4934 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689028 4934 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689033 4934 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689037 4934 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689041 4934 flags.go:64] FLAG: --enable-load-reader="false" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689046 4934 flags.go:64] FLAG: --enable-server="true" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689050 4934 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689057 4934 flags.go:64] FLAG: --event-burst="100" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689061 4934 flags.go:64] FLAG: --event-qps="50" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689065 4934 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689069 4934 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689074 4934 flags.go:64] FLAG: --eviction-hard="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689079 4934 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689083 4934 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689087 4934 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689091 4934 flags.go:64] FLAG: --eviction-soft="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689095 4934 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689099 4934 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689103 4934 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689109 4934 flags.go:64] FLAG: --experimental-mounter-path="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689113 4934 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689117 4934 flags.go:64] FLAG: --fail-swap-on="true" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689121 4934 flags.go:64] FLAG: --feature-gates="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689126 4934 flags.go:64] FLAG: --file-check-frequency="20s" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689131 4934 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689135 4934 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689139 4934 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689143 4934 flags.go:64] FLAG: --healthz-port="10248" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689148 4934 flags.go:64] FLAG: --help="false" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689152 4934 flags.go:64] FLAG: --hostname-override="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689156 4934 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689160 4934 flags.go:64] FLAG: --http-check-frequency="20s" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689165 4934 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689169 4934 flags.go:64] FLAG: --image-credential-provider-config="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689173 4934 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689177 4934 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689181 4934 flags.go:64] FLAG: --image-service-endpoint="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689185 4934 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689189 4934 flags.go:64] FLAG: --kube-api-burst="100" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689193 4934 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689197 4934 flags.go:64] FLAG: --kube-api-qps="50" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689202 4934 flags.go:64] FLAG: --kube-reserved="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689206 4934 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689210 4934 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689214 4934 flags.go:64] FLAG: --kubelet-cgroups="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689218 4934 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689222 4934 flags.go:64] FLAG: --lock-file="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689226 4934 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689230 4934 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689234 4934 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689240 4934 flags.go:64] FLAG: --log-json-split-stream="false" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689246 4934 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689250 4934 flags.go:64] FLAG: --log-text-split-stream="false" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689254 4934 flags.go:64] FLAG: --logging-format="text" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689258 4934 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689263 4934 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689267 4934 flags.go:64] FLAG: --manifest-url="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689271 4934 flags.go:64] FLAG: --manifest-url-header="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689277 4934 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689281 4934 flags.go:64] FLAG: --max-open-files="1000000" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689286 4934 flags.go:64] FLAG: --max-pods="110" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689291 4934 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689294 4934 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689299 4934 flags.go:64] FLAG: --memory-manager-policy="None" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689303 4934 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689307 4934 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689311 4934 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689315 4934 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689325 4934 flags.go:64] FLAG: --node-status-max-images="50" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689329 4934 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689333 4934 flags.go:64] FLAG: --oom-score-adj="-999" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689338 4934 flags.go:64] FLAG: --pod-cidr="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689342 4934 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689352 4934 flags.go:64] FLAG: --pod-manifest-path="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689357 4934 flags.go:64] FLAG: --pod-max-pids="-1" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689362 4934 flags.go:64] FLAG: --pods-per-core="0" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689367 4934 flags.go:64] FLAG: --port="10250" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689373 4934 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689378 4934 flags.go:64] FLAG: --provider-id="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689383 4934 flags.go:64] FLAG: --qos-reserved="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689388 4934 flags.go:64] FLAG: --read-only-port="10255" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689393 4934 flags.go:64] FLAG: --register-node="true" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689397 4934 flags.go:64] FLAG: --register-schedulable="true" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689402 4934 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689412 4934 flags.go:64] FLAG: --registry-burst="10" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689416 4934 flags.go:64] FLAG: --registry-qps="5" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689421 4934 flags.go:64] FLAG: --reserved-cpus="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689426 4934 flags.go:64] FLAG: --reserved-memory="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689432 4934 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689437 4934 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689442 4934 flags.go:64] FLAG: --rotate-certificates="false" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689446 4934 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689451 4934 flags.go:64] FLAG: --runonce="false" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689456 4934 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689461 4934 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689466 4934 flags.go:64] FLAG: --seccomp-default="false" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689471 4934 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689476 4934 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689486 4934 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689491 4934 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689497 4934 flags.go:64] FLAG: --storage-driver-password="root" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689502 4934 flags.go:64] FLAG: --storage-driver-secure="false" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689507 4934 flags.go:64] FLAG: --storage-driver-table="stats" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689512 4934 flags.go:64] FLAG: --storage-driver-user="root" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689517 4934 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689521 4934 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689526 4934 flags.go:64] FLAG: --system-cgroups="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689531 4934 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689538 4934 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689543 4934 flags.go:64] FLAG: --tls-cert-file="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689548 4934 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689554 4934 flags.go:64] FLAG: --tls-min-version="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689559 4934 flags.go:64] FLAG: --tls-private-key-file="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689565 4934 flags.go:64] FLAG: --topology-manager-policy="none" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689570 4934 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689595 4934 flags.go:64] FLAG: --topology-manager-scope="container" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689602 4934 flags.go:64] FLAG: --v="2" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689609 4934 flags.go:64] FLAG: --version="false" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689616 4934 flags.go:64] FLAG: --vmodule="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689622 4934 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.689628 4934 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689762 4934 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689770 4934 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689774 4934 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689779 4934 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689785 4934 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689791 4934 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689795 4934 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689800 4934 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689804 4934 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689809 4934 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689814 4934 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689819 4934 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689824 4934 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689830 4934 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689836 4934 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689841 4934 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689846 4934 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689850 4934 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689856 4934 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689860 4934 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689865 4934 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689870 4934 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689875 4934 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689879 4934 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689884 4934 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689889 4934 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689895 4934 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689901 4934 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689908 4934 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689913 4934 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689917 4934 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689922 4934 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689926 4934 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689930 4934 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689935 4934 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689939 4934 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689943 4934 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689947 4934 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689951 4934 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689955 4934 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689959 4934 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689965 4934 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689970 4934 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689974 4934 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689978 4934 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689982 4934 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689986 4934 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689991 4934 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689995 4934 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.689999 4934 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.690003 4934 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.690008 4934 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.690012 4934 feature_gate.go:330] unrecognized feature gate: Example Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.690017 4934 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.690022 4934 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.690027 4934 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.690032 4934 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.690036 4934 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.690041 4934 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.690046 4934 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.690050 4934 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.690054 4934 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.690059 4934 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.690063 4934 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.690068 4934 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.690072 4934 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.690077 4934 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.690081 4934 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.690086 4934 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.690090 4934 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.690094 4934 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.690109 4934 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.704388 4934 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.704447 4934 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704570 4934 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704619 4934 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704635 4934 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704649 4934 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704659 4934 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704667 4934 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704675 4934 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704683 4934 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704692 4934 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704700 4934 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704707 4934 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704715 4934 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704722 4934 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704731 4934 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704739 4934 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704747 4934 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704754 4934 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704762 4934 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704770 4934 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704778 4934 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704786 4934 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704794 4934 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704801 4934 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704810 4934 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704818 4934 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704826 4934 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704834 4934 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704842 4934 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704849 4934 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704857 4934 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704865 4934 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704872 4934 feature_gate.go:330] unrecognized feature gate: Example Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704880 4934 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704890 4934 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704901 4934 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704909 4934 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704917 4934 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704925 4934 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704932 4934 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704940 4934 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704947 4934 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704955 4934 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704963 4934 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704973 4934 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704985 4934 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.704995 4934 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705004 4934 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705013 4934 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705021 4934 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705029 4934 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705038 4934 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705046 4934 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705054 4934 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705063 4934 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705071 4934 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705081 4934 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705088 4934 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705098 4934 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705106 4934 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705114 4934 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705122 4934 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705130 4934 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705138 4934 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705148 4934 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705157 4934 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705166 4934 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705174 4934 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705182 4934 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705190 4934 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705197 4934 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705206 4934 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.705219 4934 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705439 4934 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705455 4934 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705464 4934 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705473 4934 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705483 4934 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705491 4934 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705499 4934 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705507 4934 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705516 4934 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705524 4934 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705533 4934 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705542 4934 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705551 4934 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705559 4934 feature_gate.go:330] unrecognized feature gate: Example Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705568 4934 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705606 4934 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705616 4934 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705629 4934 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705637 4934 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705646 4934 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705656 4934 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705666 4934 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705674 4934 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705683 4934 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705691 4934 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705699 4934 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705707 4934 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705715 4934 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705722 4934 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705730 4934 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705738 4934 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705746 4934 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705753 4934 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705760 4934 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705770 4934 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705778 4934 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705786 4934 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705794 4934 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705801 4934 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705809 4934 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705817 4934 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705827 4934 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705836 4934 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705844 4934 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705851 4934 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705859 4934 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705867 4934 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705875 4934 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705882 4934 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705890 4934 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705899 4934 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705906 4934 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705916 4934 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705927 4934 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705936 4934 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705944 4934 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705953 4934 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705960 4934 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705968 4934 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705978 4934 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705988 4934 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.705997 4934 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.706006 4934 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.706015 4934 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.706023 4934 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.706030 4934 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.706038 4934 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.706046 4934 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.706054 4934 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.706061 4934 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.706069 4934 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.706082 4934 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.706382 4934 server.go:940] "Client rotation is on, will bootstrap in background" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.712774 4934 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.712947 4934 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.715127 4934 server.go:997] "Starting client certificate rotation" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.715182 4934 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.715410 4934 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-21 20:40:24.509215361 +0000 UTC Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.715570 4934 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1210h51m35.793650867s for next certificate rotation Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.747863 4934 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.753832 4934 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.773882 4934 log.go:25] "Validated CRI v1 runtime API" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.804454 4934 log.go:25] "Validated CRI v1 image API" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.805935 4934 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.811356 4934 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-02-09-43-42-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.811389 4934 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.825839 4934 manager.go:217] Machine: {Timestamp:2025-10-02 09:48:48.823159789 +0000 UTC m=+0.575801331 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b BootID:cbe6531b-0669-4b46-873b-a08a7b00ffa5 Filesystems:[{Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:b3:0c:bb Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:b3:0c:bb Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:39:14:eb Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:b1:ec:b8 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:ba:c4:7c Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:71:e9:ba Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:cc:be:65 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:b2:ea:58:11:ac:4d Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:0e:be:2f:d4:c5:1c Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.826052 4934 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.826202 4934 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.828864 4934 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.829061 4934 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.829108 4934 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.829308 4934 topology_manager.go:138] "Creating topology manager with none policy" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.829318 4934 container_manager_linux.go:303] "Creating device plugin manager" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.829987 4934 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.830021 4934 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.830333 4934 state_mem.go:36] "Initialized new in-memory state store" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.830420 4934 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.837537 4934 kubelet.go:418] "Attempting to sync node with API server" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.837558 4934 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.837598 4934 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.837610 4934 kubelet.go:324] "Adding apiserver pod source" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.837623 4934 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.842122 4934 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.842845 4934 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.844660 4934 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.51:6443: connect: connection refused Oct 02 09:48:48 crc kubenswrapper[4934]: E1002 09:48:48.844762 4934 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.51:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.844958 4934 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.51:6443: connect: connection refused Oct 02 09:48:48 crc kubenswrapper[4934]: E1002 09:48:48.845087 4934 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.51:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.845715 4934 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.847406 4934 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.847432 4934 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.847440 4934 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.847448 4934 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.847461 4934 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.847470 4934 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.847479 4934 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.847492 4934 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.847503 4934 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.847513 4934 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.847531 4934 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.847539 4934 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.847568 4934 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.848079 4934 server.go:1280] "Started kubelet" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.848231 4934 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.849441 4934 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 02 09:48:48 crc systemd[1]: Started Kubernetes Kubelet. Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.857297 4934 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.51:6443: connect: connection refused Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.857869 4934 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.859569 4934 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.859649 4934 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.859832 4934 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 04:46:44.676164748 +0000 UTC Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.859862 4934 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1842h57m55.816304526s for next certificate rotation Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.859919 4934 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.859927 4934 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 02 09:48:48 crc kubenswrapper[4934]: E1002 09:48:48.859984 4934 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.861054 4934 factory.go:55] Registering systemd factory Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.861084 4934 factory.go:221] Registration of the systemd container factory successfully Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.864296 4934 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 02 09:48:48 crc kubenswrapper[4934]: E1002 09:48:48.865468 4934 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.51:6443: connect: connection refused" interval="200ms" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.865527 4934 factory.go:153] Registering CRI-O factory Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.865558 4934 factory.go:221] Registration of the crio container factory successfully Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.865621 4934 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.51:6443: connect: connection refused Oct 02 09:48:48 crc kubenswrapper[4934]: E1002 09:48:48.865688 4934 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.51:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.865659 4934 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.865727 4934 factory.go:103] Registering Raw factory Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.865775 4934 manager.go:1196] Started watching for new ooms in manager Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.866254 4934 manager.go:319] Starting recovery of all containers Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.866547 4934 server.go:460] "Adding debug handlers to kubelet server" Oct 02 09:48:48 crc kubenswrapper[4934]: E1002 09:48:48.867695 4934 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.51:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186aa3a3ea551e18 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-02 09:48:48.848043544 +0000 UTC m=+0.600685076,LastTimestamp:2025-10-02 09:48:48.848043544 +0000 UTC m=+0.600685076,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889317 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889374 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889388 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889402 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889414 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889425 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889435 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889445 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889456 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889466 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889477 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889487 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889496 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889508 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889518 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889527 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889537 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889590 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889606 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889617 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889626 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889635 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889645 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889655 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889668 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889680 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889697 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889712 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889722 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889732 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889748 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889758 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889768 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889777 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889785 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889793 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889803 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889813 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889823 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889831 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889839 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889849 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889860 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889869 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889882 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889895 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889908 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889920 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889932 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889946 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889960 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889974 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.889994 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890009 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890025 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890041 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890055 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890069 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890083 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890097 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890109 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890123 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890138 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890150 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890163 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890176 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890188 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890202 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890214 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890225 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890244 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890254 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890263 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890273 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890283 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890293 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890320 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890329 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890338 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890346 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890356 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890366 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890374 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890384 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890392 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890402 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890412 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890422 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890432 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890443 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890452 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890461 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890470 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890481 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890492 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890504 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890516 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890527 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890535 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890545 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890555 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890565 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890642 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890652 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890666 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890678 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890688 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890699 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890709 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890737 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890747 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890757 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890767 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890807 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890818 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890829 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890837 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890847 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890856 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890865 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890875 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890887 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890897 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890908 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890918 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890928 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890937 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890948 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890959 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890969 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890980 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.890991 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.891001 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.891011 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.891022 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.891034 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.891043 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.891058 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.891068 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.891078 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.891089 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.891102 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.891116 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.891130 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.891141 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.891154 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894460 4934 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894516 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894536 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894550 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894562 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894592 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894611 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894626 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894641 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894654 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894668 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894681 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894695 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894708 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894724 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894740 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894755 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894769 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894781 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894795 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894810 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894826 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894843 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894861 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894877 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894888 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894900 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894912 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894924 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894937 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894949 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894960 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894971 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894982 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.894995 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895008 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895019 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895063 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895076 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895091 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895104 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895117 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895128 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895141 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895153 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895164 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895188 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895200 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895211 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895224 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895235 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895249 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895263 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895276 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895288 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895302 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895314 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895326 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895337 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895348 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895360 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895373 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895383 4934 reconstruct.go:97] "Volume reconstruction finished" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.895390 4934 reconciler.go:26] "Reconciler: start to sync state" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.903709 4934 manager.go:324] Recovery completed Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.910126 4934 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.911828 4934 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.911869 4934 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.911924 4934 kubelet.go:2335] "Starting kubelet main sync loop" Oct 02 09:48:48 crc kubenswrapper[4934]: E1002 09:48:48.912015 4934 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 02 09:48:48 crc kubenswrapper[4934]: W1002 09:48:48.912843 4934 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.51:6443: connect: connection refused Oct 02 09:48:48 crc kubenswrapper[4934]: E1002 09:48:48.912920 4934 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.51:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.918271 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.920528 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.920611 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.920631 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.922517 4934 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.922536 4934 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.922559 4934 state_mem.go:36] "Initialized new in-memory state store" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.951532 4934 policy_none.go:49] "None policy: Start" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.952952 4934 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 02 09:48:48 crc kubenswrapper[4934]: I1002 09:48:48.952991 4934 state_mem.go:35] "Initializing new in-memory state store" Oct 02 09:48:48 crc kubenswrapper[4934]: E1002 09:48:48.960685 4934 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 09:48:49 crc kubenswrapper[4934]: E1002 09:48:49.012983 4934 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.029963 4934 manager.go:334] "Starting Device Plugin manager" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.030035 4934 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.030057 4934 server.go:79] "Starting device plugin registration server" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.030730 4934 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.030759 4934 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.030955 4934 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.031148 4934 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.031167 4934 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 02 09:48:49 crc kubenswrapper[4934]: E1002 09:48:49.040747 4934 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 02 09:48:49 crc kubenswrapper[4934]: E1002 09:48:49.066717 4934 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.51:6443: connect: connection refused" interval="400ms" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.131693 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.134227 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.134296 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.134319 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.134366 4934 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 09:48:49 crc kubenswrapper[4934]: E1002 09:48:49.135090 4934 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.51:6443: connect: connection refused" node="crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.214109 4934 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc"] Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.214262 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.215317 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.215346 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.215354 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.215477 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.215820 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.215891 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.216147 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.216193 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.216209 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.216321 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.216439 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.216476 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.216979 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.217014 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.217029 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.217223 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.217236 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.217245 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.217425 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.217485 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.217506 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.217787 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.217898 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.217936 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.218831 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.218844 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.218875 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.218888 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.218932 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.218901 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.219088 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.219143 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.219160 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.219955 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.219980 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.219990 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.220420 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.220471 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.220485 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.220764 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.220800 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.221719 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.221759 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.221776 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.300658 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.300763 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.300817 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.300842 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.300863 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.300908 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.300927 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.300945 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.300987 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.301006 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.301027 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.301069 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.301091 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.301113 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.301157 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.335999 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.337344 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.337397 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.337512 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.337545 4934 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 09:48:49 crc kubenswrapper[4934]: E1002 09:48:49.338429 4934 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.51:6443: connect: connection refused" node="crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.402280 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.402387 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.402464 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.402507 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.402546 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.402557 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.402639 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.402650 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.402687 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.402721 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.402729 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.402745 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.402773 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.402791 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.402805 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.402827 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.402822 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.402846 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.402911 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.402874 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.402784 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.402986 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.402984 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.403067 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.403037 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.403029 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.403187 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.403240 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.403348 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.403448 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: E1002 09:48:49.468416 4934 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.51:6443: connect: connection refused" interval="800ms" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.544751 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.551951 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.572437 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.580035 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.585537 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 02 09:48:49 crc kubenswrapper[4934]: W1002 09:48:49.591820 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-39dac8511532c689ee210ace5221fcdb5d91fbc8812b6cb80e54155803105b6d WatchSource:0}: Error finding container 39dac8511532c689ee210ace5221fcdb5d91fbc8812b6cb80e54155803105b6d: Status 404 returned error can't find the container with id 39dac8511532c689ee210ace5221fcdb5d91fbc8812b6cb80e54155803105b6d Oct 02 09:48:49 crc kubenswrapper[4934]: W1002 09:48:49.594464 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-eda44c769b7d2c3a5070c4d11e3a1eace11014391c231005363268feea86f064 WatchSource:0}: Error finding container eda44c769b7d2c3a5070c4d11e3a1eace11014391c231005363268feea86f064: Status 404 returned error can't find the container with id eda44c769b7d2c3a5070c4d11e3a1eace11014391c231005363268feea86f064 Oct 02 09:48:49 crc kubenswrapper[4934]: W1002 09:48:49.611406 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-2900085b8bd9bd7e0e0896a535ff8eb877763e112dda8e3ea22a64b5d282c54c WatchSource:0}: Error finding container 2900085b8bd9bd7e0e0896a535ff8eb877763e112dda8e3ea22a64b5d282c54c: Status 404 returned error can't find the container with id 2900085b8bd9bd7e0e0896a535ff8eb877763e112dda8e3ea22a64b5d282c54c Oct 02 09:48:49 crc kubenswrapper[4934]: W1002 09:48:49.612865 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-eee9d0e59593789baa8ddba2c8e8833a78d6fca764d6a2cce3d286d35f7dbcf2 WatchSource:0}: Error finding container eee9d0e59593789baa8ddba2c8e8833a78d6fca764d6a2cce3d286d35f7dbcf2: Status 404 returned error can't find the container with id eee9d0e59593789baa8ddba2c8e8833a78d6fca764d6a2cce3d286d35f7dbcf2 Oct 02 09:48:49 crc kubenswrapper[4934]: W1002 09:48:49.725141 4934 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.51:6443: connect: connection refused Oct 02 09:48:49 crc kubenswrapper[4934]: E1002 09:48:49.725254 4934 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.51:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.739592 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.741013 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.741048 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.741059 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.741082 4934 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 09:48:49 crc kubenswrapper[4934]: E1002 09:48:49.741569 4934 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.51:6443: connect: connection refused" node="crc" Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.858295 4934 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.51:6443: connect: connection refused Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.916392 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"eda44c769b7d2c3a5070c4d11e3a1eace11014391c231005363268feea86f064"} Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.918504 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"39dac8511532c689ee210ace5221fcdb5d91fbc8812b6cb80e54155803105b6d"} Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.920823 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"eee9d0e59593789baa8ddba2c8e8833a78d6fca764d6a2cce3d286d35f7dbcf2"} Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.921491 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"2900085b8bd9bd7e0e0896a535ff8eb877763e112dda8e3ea22a64b5d282c54c"} Oct 02 09:48:49 crc kubenswrapper[4934]: I1002 09:48:49.922900 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"395dab66271885ccea8af370fb9f0a0d70757787d6805d96b6cdf74b5e5d2fa0"} Oct 02 09:48:50 crc kubenswrapper[4934]: W1002 09:48:50.077149 4934 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.51:6443: connect: connection refused Oct 02 09:48:50 crc kubenswrapper[4934]: E1002 09:48:50.077234 4934 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.51:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:48:50 crc kubenswrapper[4934]: W1002 09:48:50.251443 4934 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.51:6443: connect: connection refused Oct 02 09:48:50 crc kubenswrapper[4934]: E1002 09:48:50.251795 4934 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.51:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:48:50 crc kubenswrapper[4934]: E1002 09:48:50.269944 4934 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.51:6443: connect: connection refused" interval="1.6s" Oct 02 09:48:50 crc kubenswrapper[4934]: W1002 09:48:50.492230 4934 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.51:6443: connect: connection refused Oct 02 09:48:50 crc kubenswrapper[4934]: E1002 09:48:50.492348 4934 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.51:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.542535 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.544379 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.544431 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.544442 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.544475 4934 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 09:48:50 crc kubenswrapper[4934]: E1002 09:48:50.545024 4934 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.51:6443: connect: connection refused" node="crc" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.859012 4934 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.51:6443: connect: connection refused Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.927647 4934 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff" exitCode=0 Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.927778 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.928133 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff"} Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.929076 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.929126 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.929144 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.929564 4934 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="1f0b4710144b27bd0899de27f526502d6b23a51bcffa2ecd7a41efd245fff5da" exitCode=0 Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.929736 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.930141 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"1f0b4710144b27bd0899de27f526502d6b23a51bcffa2ecd7a41efd245fff5da"} Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.930991 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.931030 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.931044 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.932680 4934 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="0c6711e8eee36c5d872a158ccedc38e3739b94e96e19564babf96bd6fef2dad0" exitCode=0 Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.932764 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.932769 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"0c6711e8eee36c5d872a158ccedc38e3739b94e96e19564babf96bd6fef2dad0"} Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.934463 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.934507 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.934517 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.935692 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd"} Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.935725 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.935740 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f"} Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.935761 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4"} Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.935774 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182"} Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.936469 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.936501 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.936514 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.937440 4934 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47" exitCode=0 Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.937687 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47"} Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.937757 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.938535 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.938587 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.938602 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.940872 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.942104 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.942140 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:50 crc kubenswrapper[4934]: I1002 09:48:50.942153 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.858530 4934 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.51:6443: connect: connection refused Oct 02 09:48:51 crc kubenswrapper[4934]: E1002 09:48:51.871328 4934 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.51:6443: connect: connection refused" interval="3.2s" Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.941085 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.941080 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"ab08caf3ec6f4cec18fc54615059138c4110223b1413c5fbbc1cf718dade62f4"} Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.941891 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.941922 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.941931 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.943823 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9af6ae65b40a9786691a168f58a28aa4e1f1fdb401976046e4657bbbc0d44fc2"} Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.943858 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"4f14dcea7cf5a10dfe10430bfd3d6e41710d33685f43cdc634206a17bbb3697f"} Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.943869 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"407acc937940ce03f746be72752c5efa2ce7e416c39a80c2763befdec18daa56"} Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.943884 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.945086 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.945117 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.945131 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.948665 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f950f5b95e378c306df980ebec95b90b817a0720a8582a2b12ef1533f71e43b6"} Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.948701 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2"} Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.948715 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c"} Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.948732 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd"} Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.948745 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca"} Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.948734 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.949547 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.949594 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.949605 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.951036 4934 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec" exitCode=0 Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.951095 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.951121 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec"} Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.951172 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.951675 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.951698 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.951708 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.951938 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.951972 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:51 crc kubenswrapper[4934]: I1002 09:48:51.951984 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:52 crc kubenswrapper[4934]: I1002 09:48:52.145448 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:52 crc kubenswrapper[4934]: I1002 09:48:52.146677 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:52 crc kubenswrapper[4934]: I1002 09:48:52.146715 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:52 crc kubenswrapper[4934]: I1002 09:48:52.146725 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:52 crc kubenswrapper[4934]: I1002 09:48:52.146750 4934 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 09:48:52 crc kubenswrapper[4934]: E1002 09:48:52.147197 4934 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.51:6443: connect: connection refused" node="crc" Oct 02 09:48:52 crc kubenswrapper[4934]: I1002 09:48:52.954989 4934 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34" exitCode=0 Oct 02 09:48:52 crc kubenswrapper[4934]: I1002 09:48:52.955113 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:52 crc kubenswrapper[4934]: I1002 09:48:52.955125 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34"} Oct 02 09:48:52 crc kubenswrapper[4934]: I1002 09:48:52.955140 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:52 crc kubenswrapper[4934]: I1002 09:48:52.955170 4934 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 09:48:52 crc kubenswrapper[4934]: I1002 09:48:52.955113 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:52 crc kubenswrapper[4934]: I1002 09:48:52.955229 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:52 crc kubenswrapper[4934]: I1002 09:48:52.955187 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:48:52 crc kubenswrapper[4934]: I1002 09:48:52.956220 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:52 crc kubenswrapper[4934]: I1002 09:48:52.956244 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:52 crc kubenswrapper[4934]: I1002 09:48:52.956253 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:52 crc kubenswrapper[4934]: I1002 09:48:52.956854 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:52 crc kubenswrapper[4934]: I1002 09:48:52.956871 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:52 crc kubenswrapper[4934]: I1002 09:48:52.956879 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:52 crc kubenswrapper[4934]: I1002 09:48:52.956975 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:52 crc kubenswrapper[4934]: I1002 09:48:52.956997 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:52 crc kubenswrapper[4934]: I1002 09:48:52.957008 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:52 crc kubenswrapper[4934]: I1002 09:48:52.956979 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:52 crc kubenswrapper[4934]: I1002 09:48:52.957064 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:52 crc kubenswrapper[4934]: I1002 09:48:52.957080 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:53 crc kubenswrapper[4934]: I1002 09:48:53.961306 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51"} Oct 02 09:48:53 crc kubenswrapper[4934]: I1002 09:48:53.961361 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:53 crc kubenswrapper[4934]: I1002 09:48:53.961371 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3"} Oct 02 09:48:53 crc kubenswrapper[4934]: I1002 09:48:53.961394 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b"} Oct 02 09:48:53 crc kubenswrapper[4934]: I1002 09:48:53.961412 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a"} Oct 02 09:48:53 crc kubenswrapper[4934]: I1002 09:48:53.961429 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566"} Oct 02 09:48:53 crc kubenswrapper[4934]: I1002 09:48:53.961935 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:53 crc kubenswrapper[4934]: I1002 09:48:53.962273 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:53 crc kubenswrapper[4934]: I1002 09:48:53.962297 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:53 crc kubenswrapper[4934]: I1002 09:48:53.962305 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:53 crc kubenswrapper[4934]: I1002 09:48:53.962932 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:53 crc kubenswrapper[4934]: I1002 09:48:53.962961 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:53 crc kubenswrapper[4934]: I1002 09:48:53.962974 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:54 crc kubenswrapper[4934]: I1002 09:48:54.111173 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:48:54 crc kubenswrapper[4934]: I1002 09:48:54.964130 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:54 crc kubenswrapper[4934]: I1002 09:48:54.964157 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:54 crc kubenswrapper[4934]: I1002 09:48:54.965655 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:54 crc kubenswrapper[4934]: I1002 09:48:54.965703 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:54 crc kubenswrapper[4934]: I1002 09:48:54.965717 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:54 crc kubenswrapper[4934]: I1002 09:48:54.965737 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:54 crc kubenswrapper[4934]: I1002 09:48:54.965787 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:54 crc kubenswrapper[4934]: I1002 09:48:54.965808 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:55 crc kubenswrapper[4934]: I1002 09:48:55.018254 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:48:55 crc kubenswrapper[4934]: I1002 09:48:55.348280 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:55 crc kubenswrapper[4934]: I1002 09:48:55.349895 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:55 crc kubenswrapper[4934]: I1002 09:48:55.350040 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:55 crc kubenswrapper[4934]: I1002 09:48:55.350081 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:55 crc kubenswrapper[4934]: I1002 09:48:55.350128 4934 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 09:48:55 crc kubenswrapper[4934]: I1002 09:48:55.558924 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:48:55 crc kubenswrapper[4934]: I1002 09:48:55.559105 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:55 crc kubenswrapper[4934]: I1002 09:48:55.560283 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:55 crc kubenswrapper[4934]: I1002 09:48:55.560325 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:55 crc kubenswrapper[4934]: I1002 09:48:55.560336 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:55 crc kubenswrapper[4934]: I1002 09:48:55.967112 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:55 crc kubenswrapper[4934]: I1002 09:48:55.968306 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:55 crc kubenswrapper[4934]: I1002 09:48:55.968340 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:55 crc kubenswrapper[4934]: I1002 09:48:55.968348 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:57 crc kubenswrapper[4934]: I1002 09:48:57.247027 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:48:57 crc kubenswrapper[4934]: I1002 09:48:57.247410 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:57 crc kubenswrapper[4934]: I1002 09:48:57.248974 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:57 crc kubenswrapper[4934]: I1002 09:48:57.249026 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:57 crc kubenswrapper[4934]: I1002 09:48:57.249042 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:58 crc kubenswrapper[4934]: I1002 09:48:58.800631 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 02 09:48:58 crc kubenswrapper[4934]: I1002 09:48:58.800943 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:58 crc kubenswrapper[4934]: I1002 09:48:58.803124 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:58 crc kubenswrapper[4934]: I1002 09:48:58.803198 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:58 crc kubenswrapper[4934]: I1002 09:48:58.803224 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:58 crc kubenswrapper[4934]: I1002 09:48:58.959624 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:48:58 crc kubenswrapper[4934]: I1002 09:48:58.959866 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:58 crc kubenswrapper[4934]: I1002 09:48:58.961402 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:58 crc kubenswrapper[4934]: I1002 09:48:58.961466 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:58 crc kubenswrapper[4934]: I1002 09:48:58.961508 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:59 crc kubenswrapper[4934]: E1002 09:48:59.040897 4934 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 02 09:48:59 crc kubenswrapper[4934]: I1002 09:48:59.706369 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:48:59 crc kubenswrapper[4934]: I1002 09:48:59.706547 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:59 crc kubenswrapper[4934]: I1002 09:48:59.708134 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:59 crc kubenswrapper[4934]: I1002 09:48:59.708214 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:59 crc kubenswrapper[4934]: I1002 09:48:59.708228 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:48:59 crc kubenswrapper[4934]: I1002 09:48:59.711390 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:48:59 crc kubenswrapper[4934]: I1002 09:48:59.976617 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:48:59 crc kubenswrapper[4934]: I1002 09:48:59.977779 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:48:59 crc kubenswrapper[4934]: I1002 09:48:59.977850 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:48:59 crc kubenswrapper[4934]: I1002 09:48:59.977886 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:00 crc kubenswrapper[4934]: I1002 09:49:00.248084 4934 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 02 09:49:00 crc kubenswrapper[4934]: I1002 09:49:00.248201 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 09:49:00 crc kubenswrapper[4934]: I1002 09:49:00.322876 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 02 09:49:00 crc kubenswrapper[4934]: I1002 09:49:00.323206 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:49:00 crc kubenswrapper[4934]: I1002 09:49:00.324932 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:00 crc kubenswrapper[4934]: I1002 09:49:00.324986 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:00 crc kubenswrapper[4934]: I1002 09:49:00.325009 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:00 crc kubenswrapper[4934]: I1002 09:49:00.563932 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:49:00 crc kubenswrapper[4934]: I1002 09:49:00.980364 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:49:00 crc kubenswrapper[4934]: I1002 09:49:00.981787 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:00 crc kubenswrapper[4934]: I1002 09:49:00.981863 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:00 crc kubenswrapper[4934]: I1002 09:49:00.981878 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:00 crc kubenswrapper[4934]: I1002 09:49:00.986979 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:49:01 crc kubenswrapper[4934]: I1002 09:49:01.983713 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:49:01 crc kubenswrapper[4934]: I1002 09:49:01.985024 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:01 crc kubenswrapper[4934]: I1002 09:49:01.985063 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:01 crc kubenswrapper[4934]: I1002 09:49:01.985080 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:02 crc kubenswrapper[4934]: W1002 09:49:02.642319 4934 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 02 09:49:02 crc kubenswrapper[4934]: I1002 09:49:02.642668 4934 trace.go:236] Trace[2136953637]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 09:48:52.640) (total time: 10002ms): Oct 02 09:49:02 crc kubenswrapper[4934]: Trace[2136953637]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (09:49:02.642) Oct 02 09:49:02 crc kubenswrapper[4934]: Trace[2136953637]: [10.002024944s] [10.002024944s] END Oct 02 09:49:02 crc kubenswrapper[4934]: E1002 09:49:02.642808 4934 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 02 09:49:02 crc kubenswrapper[4934]: W1002 09:49:02.734652 4934 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 02 09:49:02 crc kubenswrapper[4934]: I1002 09:49:02.735023 4934 trace.go:236] Trace[915476683]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 09:48:52.732) (total time: 10002ms): Oct 02 09:49:02 crc kubenswrapper[4934]: Trace[915476683]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (09:49:02.734) Oct 02 09:49:02 crc kubenswrapper[4934]: Trace[915476683]: [10.002213469s] [10.002213469s] END Oct 02 09:49:02 crc kubenswrapper[4934]: E1002 09:49:02.735345 4934 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 02 09:49:02 crc kubenswrapper[4934]: I1002 09:49:02.766100 4934 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:52166->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 02 09:49:02 crc kubenswrapper[4934]: I1002 09:49:02.766152 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:52166->192.168.126.11:17697: read: connection reset by peer" Oct 02 09:49:02 crc kubenswrapper[4934]: I1002 09:49:02.859408 4934 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 02 09:49:02 crc kubenswrapper[4934]: W1002 09:49:02.891139 4934 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 02 09:49:02 crc kubenswrapper[4934]: I1002 09:49:02.891242 4934 trace.go:236] Trace[1486218560]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 09:48:52.889) (total time: 10001ms): Oct 02 09:49:02 crc kubenswrapper[4934]: Trace[1486218560]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (09:49:02.891) Oct 02 09:49:02 crc kubenswrapper[4934]: Trace[1486218560]: [10.001673747s] [10.001673747s] END Oct 02 09:49:02 crc kubenswrapper[4934]: E1002 09:49:02.891267 4934 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 02 09:49:02 crc kubenswrapper[4934]: I1002 09:49:02.989404 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 02 09:49:02 crc kubenswrapper[4934]: I1002 09:49:02.993075 4934 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f950f5b95e378c306df980ebec95b90b817a0720a8582a2b12ef1533f71e43b6" exitCode=255 Oct 02 09:49:02 crc kubenswrapper[4934]: I1002 09:49:02.993153 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"f950f5b95e378c306df980ebec95b90b817a0720a8582a2b12ef1533f71e43b6"} Oct 02 09:49:02 crc kubenswrapper[4934]: I1002 09:49:02.993428 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:49:02 crc kubenswrapper[4934]: I1002 09:49:02.994439 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:02 crc kubenswrapper[4934]: I1002 09:49:02.994477 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:02 crc kubenswrapper[4934]: I1002 09:49:02.994494 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:02 crc kubenswrapper[4934]: I1002 09:49:02.995213 4934 scope.go:117] "RemoveContainer" containerID="f950f5b95e378c306df980ebec95b90b817a0720a8582a2b12ef1533f71e43b6" Oct 02 09:49:03 crc kubenswrapper[4934]: I1002 09:49:03.135725 4934 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 02 09:49:03 crc kubenswrapper[4934]: I1002 09:49:03.135866 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 02 09:49:03 crc kubenswrapper[4934]: I1002 09:49:03.141254 4934 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 02 09:49:03 crc kubenswrapper[4934]: I1002 09:49:03.141309 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 02 09:49:03 crc kubenswrapper[4934]: I1002 09:49:03.997097 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 02 09:49:03 crc kubenswrapper[4934]: I1002 09:49:03.998874 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97"} Oct 02 09:49:03 crc kubenswrapper[4934]: I1002 09:49:03.999013 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:49:04 crc kubenswrapper[4934]: I1002 09:49:03.999947 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:04 crc kubenswrapper[4934]: I1002 09:49:03.999998 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:04 crc kubenswrapper[4934]: I1002 09:49:04.000016 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:05 crc kubenswrapper[4934]: I1002 09:49:05.025217 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:49:05 crc kubenswrapper[4934]: I1002 09:49:05.025353 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:49:05 crc kubenswrapper[4934]: I1002 09:49:05.025481 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:49:05 crc kubenswrapper[4934]: I1002 09:49:05.026229 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:05 crc kubenswrapper[4934]: I1002 09:49:05.026255 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:05 crc kubenswrapper[4934]: I1002 09:49:05.026264 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:05 crc kubenswrapper[4934]: I1002 09:49:05.030523 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:49:06 crc kubenswrapper[4934]: I1002 09:49:06.003630 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:49:06 crc kubenswrapper[4934]: I1002 09:49:06.004509 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:06 crc kubenswrapper[4934]: I1002 09:49:06.004563 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:06 crc kubenswrapper[4934]: I1002 09:49:06.004617 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:06 crc kubenswrapper[4934]: I1002 09:49:06.675636 4934 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.005608 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.008084 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.008167 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.008193 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.625389 4934 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.852293 4934 apiserver.go:52] "Watching apiserver" Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.856292 4934 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.856772 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.857254 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.857396 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:07 crc kubenswrapper[4934]: E1002 09:49:07.857443 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.857552 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:07 crc kubenswrapper[4934]: E1002 09:49:07.857713 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.857981 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.858053 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:07 crc kubenswrapper[4934]: E1002 09:49:07.858141 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.858211 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.859366 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.859871 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.860271 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.860315 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.860284 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.860453 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.860568 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.860823 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.860862 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.865335 4934 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.893516 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.903779 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.913906 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.919370 4934 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.924304 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.934289 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.945072 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.953425 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:07 crc kubenswrapper[4934]: I1002 09:49:07.962878 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:08 crc kubenswrapper[4934]: E1002 09:49:08.124941 4934 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.127132 4934 trace.go:236] Trace[1296260062]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 09:48:53.480) (total time: 14646ms): Oct 02 09:49:08 crc kubenswrapper[4934]: Trace[1296260062]: ---"Objects listed" error: 14646ms (09:49:08.127) Oct 02 09:49:08 crc kubenswrapper[4934]: Trace[1296260062]: [14.646224565s] [14.646224565s] END Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.127347 4934 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.128374 4934 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 02 09:49:08 crc kubenswrapper[4934]: E1002 09:49:08.129085 4934 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.229455 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.229773 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.229900 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.230017 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.230143 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.230263 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.230382 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.230033 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.230860 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.230193 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.230934 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.230324 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.230472 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.230963 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.230984 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231005 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.230482 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231025 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231042 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.230627 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.230636 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231060 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231080 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231135 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231161 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231198 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231215 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231237 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231274 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231290 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231305 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231321 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231336 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231351 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231394 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231411 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231421 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231449 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231476 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231661 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231696 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231730 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231428 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231766 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231784 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231802 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231817 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231820 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.233410 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.231867 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.233503 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.233537 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.233567 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.233614 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.233645 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.233673 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.233697 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.233724 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.233751 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.233778 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.233809 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.233837 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.233864 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.233887 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.233918 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.233947 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.233974 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234004 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234026 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234055 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234080 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234103 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234128 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234151 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234157 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234178 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234259 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234260 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234288 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234321 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234339 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234359 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234380 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234397 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234418 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234440 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234487 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234505 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234533 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234620 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234638 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234659 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234681 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234700 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234720 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234739 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234760 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234777 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234797 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234815 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234833 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234852 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234877 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234893 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234914 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234935 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234951 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234971 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234991 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235009 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235025 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235044 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235065 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235084 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235105 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235123 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235140 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235158 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235179 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235201 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235220 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235239 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235259 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235276 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235296 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235316 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235338 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235355 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235376 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235396 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235414 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235435 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235455 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235477 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235497 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235515 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235541 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235559 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235594 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235617 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235634 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235655 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235676 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235696 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235717 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235737 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235756 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235772 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235801 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235822 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235840 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235861 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235880 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235900 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235977 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236003 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236020 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236038 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236058 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236075 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236096 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236116 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236137 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236154 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236173 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236194 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236210 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236228 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236249 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236268 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236292 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236312 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236327 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236347 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236365 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236382 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236402 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236420 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236438 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236457 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236475 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236493 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236510 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236530 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236886 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236907 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236928 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236950 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236971 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236991 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237013 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237036 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237055 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237076 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237098 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237123 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237145 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237164 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237184 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237202 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237225 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237249 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237273 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237300 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237512 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237539 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237562 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237599 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237621 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237639 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237661 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237680 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237740 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237766 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237796 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237821 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237846 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237869 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237889 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237907 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237926 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237948 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237969 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237993 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.238016 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.238038 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.238132 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.238146 4934 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.238161 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.238173 4934 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.238184 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.238195 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.238209 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.238308 4934 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.238321 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.238331 4934 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.238345 4934 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.238356 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.238366 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.238379 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.238389 4934 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.238401 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.238412 4934 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234415 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234649 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234684 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234783 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.234964 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.240492 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235047 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235156 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235164 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235301 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235340 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235609 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235670 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235775 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.235932 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236021 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236120 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236370 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236513 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236557 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236674 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.236812 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237009 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237072 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237219 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237324 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237310 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237392 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237388 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237373 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237453 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237656 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237562 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.237719 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.238118 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.238148 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.238519 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: E1002 09:49:08.238808 4934 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.239224 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.239236 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.239472 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.239543 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.239641 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.239874 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.240411 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.240425 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.240435 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.240479 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.240823 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.240993 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.241046 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.241084 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.241157 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.241301 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.241340 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.241324 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.241435 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.241459 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.241468 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.241523 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.241535 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: E1002 09:49:08.242138 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:49:08.742102792 +0000 UTC m=+20.494744314 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.242892 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.243039 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.243184 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.243223 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.243233 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.243398 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.243508 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.243622 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.243821 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.243833 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.244120 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.244324 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.244362 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.244608 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.244549 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.245026 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.244709 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.245394 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.245487 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.241481 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.245743 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.245084 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.245805 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.247168 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.247396 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.247485 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.247593 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.247615 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.247841 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.247969 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.247934 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.248207 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: E1002 09:49:08.248679 4934 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:49:08 crc kubenswrapper[4934]: E1002 09:49:08.248756 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:49:08.748735826 +0000 UTC m=+20.501377348 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.249478 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.249699 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.249765 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.249786 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.249912 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.249925 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.245519 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.250043 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: E1002 09:49:08.250224 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:49:08.750130318 +0000 UTC m=+20.502772020 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.250105 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.250275 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.250399 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.250439 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.250594 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.251764 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.251843 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.252036 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.252160 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.252279 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.252285 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.252552 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.252591 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.252811 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.252958 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.253044 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.253123 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.253302 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.254180 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.256220 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.256495 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.256783 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.256381 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.257346 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.257701 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.257912 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.257892 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.258280 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.258536 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.259037 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.259248 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.259489 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.259893 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.260000 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.260328 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.260811 4934 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.260811 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.260975 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.261373 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.261401 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.261418 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.245842 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.245888 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.262111 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.246189 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.262360 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.262627 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.262926 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.264910 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.265224 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: E1002 09:49:08.265569 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:49:08 crc kubenswrapper[4934]: E1002 09:49:08.265620 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:49:08 crc kubenswrapper[4934]: E1002 09:49:08.265640 4934 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:49:08 crc kubenswrapper[4934]: E1002 09:49:08.266814 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 09:49:08.765719238 +0000 UTC m=+20.518360960 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:49:08 crc kubenswrapper[4934]: E1002 09:49:08.267446 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:49:08 crc kubenswrapper[4934]: E1002 09:49:08.267472 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:49:08 crc kubenswrapper[4934]: E1002 09:49:08.267486 4934 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:49:08 crc kubenswrapper[4934]: E1002 09:49:08.267561 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 09:49:08.767544789 +0000 UTC m=+20.520186511 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.272764 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.272785 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.272984 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.273076 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.273218 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.275282 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.275657 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.276510 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.276724 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.276835 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.276943 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.277118 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.277303 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.276369 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.277687 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.278158 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.279091 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.283157 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.283258 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.285912 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.286159 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.286770 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.286868 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.286918 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.287023 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.287432 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.287497 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.287637 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.287951 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.293423 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.293491 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.293568 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.295230 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.312689 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.317843 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.319266 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339389 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339448 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339490 4934 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339501 4934 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339514 4934 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339523 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339533 4934 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339542 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339551 4934 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339561 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339570 4934 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339593 4934 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339602 4934 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339610 4934 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339619 4934 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339627 4934 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339635 4934 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339646 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339655 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339664 4934 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339673 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339681 4934 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339689 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339697 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339705 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339713 4934 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339721 4934 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339728 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339736 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339744 4934 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339752 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339760 4934 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339767 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339775 4934 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339782 4934 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339791 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339799 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339807 4934 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339817 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339826 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339834 4934 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339843 4934 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339855 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339863 4934 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339873 4934 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339881 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339890 4934 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339899 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339907 4934 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339917 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339926 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339935 4934 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339944 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339952 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339960 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339968 4934 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339976 4934 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.339984 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340007 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340016 4934 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340025 4934 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340041 4934 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340049 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340059 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340068 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340076 4934 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340085 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340093 4934 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340102 4934 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340110 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340120 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340128 4934 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340135 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340143 4934 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340153 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340162 4934 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340169 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340178 4934 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340185 4934 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340193 4934 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340201 4934 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340209 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340236 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340296 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340306 4934 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340314 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340322 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340298 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340332 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340450 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340494 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340510 4934 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340525 4934 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340541 4934 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340555 4934 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340570 4934 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340610 4934 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340624 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340640 4934 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340655 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340674 4934 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340689 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340704 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340718 4934 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340734 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340751 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340768 4934 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340783 4934 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340798 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340813 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340829 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340844 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340860 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340876 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340889 4934 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340902 4934 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340916 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340957 4934 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340976 4934 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.340990 4934 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341004 4934 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341018 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341032 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341045 4934 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341059 4934 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341073 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341087 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341100 4934 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341114 4934 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341141 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341156 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341171 4934 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341188 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341203 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341217 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341230 4934 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341244 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341258 4934 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341272 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341286 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341300 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341323 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341336 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341349 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341363 4934 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341377 4934 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341391 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341405 4934 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341433 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341447 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341465 4934 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341478 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341494 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341508 4934 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341520 4934 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341535 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341548 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341562 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341603 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341618 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341628 4934 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341638 4934 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341649 4934 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341659 4934 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341668 4934 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341679 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341689 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341698 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341713 4934 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341729 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341743 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341755 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341766 4934 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341778 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341791 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341804 4934 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341814 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341829 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341841 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341854 4934 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.341887 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.470798 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.480098 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.486652 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.590353 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.594802 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.601911 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.608195 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.618111 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.629878 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.645319 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.657327 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.666620 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.683800 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.697830 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.707217 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.717367 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.727918 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.740266 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.746260 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:08 crc kubenswrapper[4934]: E1002 09:49:08.746502 4934 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:49:08 crc kubenswrapper[4934]: E1002 09:49:08.746566 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:49:09.746548924 +0000 UTC m=+21.499190446 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.750821 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.849463 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.849655 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.849758 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:08 crc kubenswrapper[4934]: E1002 09:49:08.849849 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:49:09.849824877 +0000 UTC m=+21.602466399 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.849911 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:08 crc kubenswrapper[4934]: E1002 09:49:08.850005 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:49:08 crc kubenswrapper[4934]: E1002 09:49:08.850033 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:49:08 crc kubenswrapper[4934]: E1002 09:49:08.850044 4934 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:49:08 crc kubenswrapper[4934]: E1002 09:49:08.850070 4934 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:49:08 crc kubenswrapper[4934]: E1002 09:49:08.850102 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 09:49:09.850076663 +0000 UTC m=+21.602718185 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:49:08 crc kubenswrapper[4934]: E1002 09:49:08.850131 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:49:09.850120474 +0000 UTC m=+21.602761996 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:49:08 crc kubenswrapper[4934]: E1002 09:49:08.850650 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:49:08 crc kubenswrapper[4934]: E1002 09:49:08.850676 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:49:08 crc kubenswrapper[4934]: E1002 09:49:08.850684 4934 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:49:08 crc kubenswrapper[4934]: E1002 09:49:08.850718 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 09:49:09.850704517 +0000 UTC m=+21.603346039 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.916571 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.917407 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.918458 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.919123 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.920149 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.920685 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.921400 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.922496 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.923276 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.923877 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.924358 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.924994 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.926258 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.926849 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.927358 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.928255 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.928761 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.929814 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.930295 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.930955 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.932049 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.932610 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.933831 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.934409 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.935605 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.936125 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.936710 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.937978 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.938568 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.939536 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.939674 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.940168 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.941370 4934 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.941510 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.943266 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.944231 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.944824 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.946651 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.947390 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.948404 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.949182 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.950393 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.951046 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.951678 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.952219 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.952946 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.954106 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.954663 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.955760 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.956293 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.957591 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.958090 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.958981 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.959764 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.960879 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.961520 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.961989 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.963908 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.973784 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.986632 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:08 crc kubenswrapper[4934]: I1002 09:49:08.996394 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.012462 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.013650 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.015922 4934 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97" exitCode=255 Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.015982 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97"} Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.016028 4934 scope.go:117] "RemoveContainer" containerID="f950f5b95e378c306df980ebec95b90b817a0720a8582a2b12ef1533f71e43b6" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.017136 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"40d965e9188be5c5ad166f1df1206b18585b0caed46817654282e4532a774714"} Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.019540 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e"} Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.019599 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c"} Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.019611 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"8ddbac2f9da00f1b4f880cae9c47148396f8b41e033cbb23c4d009b412b87545"} Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.021094 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df"} Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.021131 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"3107c8dc3b0529d1a256288dac116fe680d835eed7d5037e598ecc8838e28663"} Oct 02 09:49:09 crc kubenswrapper[4934]: E1002 09:49:09.029422 4934 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.034863 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.035464 4934 scope.go:117] "RemoveContainer" containerID="a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97" Oct 02 09:49:09 crc kubenswrapper[4934]: E1002 09:49:09.036293 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.038669 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.049182 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.074767 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.089614 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.102402 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.113772 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.125066 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.135884 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.147078 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.159213 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.170196 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.182132 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.199165 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.209852 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.224095 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f950f5b95e378c306df980ebec95b90b817a0720a8582a2b12ef1533f71e43b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:02Z\\\",\\\"message\\\":\\\"W1002 09:48:51.991265 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 09:48:51.991593 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759398531 cert, and key in /tmp/serving-cert-3222084787/serving-signer.crt, /tmp/serving-cert-3222084787/serving-signer.key\\\\nI1002 09:48:52.252472 1 observer_polling.go:159] Starting file observer\\\\nW1002 09:48:52.256275 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 09:48:52.256431 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:48:52.257041 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3222084787/tls.crt::/tmp/serving-cert-3222084787/tls.key\\\\\\\"\\\\nF1002 09:49:02.761346 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.757173 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:09 crc kubenswrapper[4934]: E1002 09:49:09.757318 4934 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:49:09 crc kubenswrapper[4934]: E1002 09:49:09.757395 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:49:11.757364891 +0000 UTC m=+23.510006413 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.848661 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-85mph"] Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.849041 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-85mph" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.855141 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.858006 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.858091 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.858122 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.858149 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:09 crc kubenswrapper[4934]: E1002 09:49:09.858196 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:49:11.858168518 +0000 UTC m=+23.610810040 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:49:09 crc kubenswrapper[4934]: E1002 09:49:09.858269 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:49:09 crc kubenswrapper[4934]: E1002 09:49:09.858296 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:49:09 crc kubenswrapper[4934]: E1002 09:49:09.858308 4934 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:49:09 crc kubenswrapper[4934]: E1002 09:49:09.858270 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:49:09 crc kubenswrapper[4934]: E1002 09:49:09.858334 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:49:09 crc kubenswrapper[4934]: E1002 09:49:09.858356 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 09:49:11.858339561 +0000 UTC m=+23.610981083 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:49:09 crc kubenswrapper[4934]: E1002 09:49:09.858359 4934 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:49:09 crc kubenswrapper[4934]: E1002 09:49:09.858388 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 09:49:11.858381342 +0000 UTC m=+23.611022954 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:49:09 crc kubenswrapper[4934]: E1002 09:49:09.858273 4934 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:49:09 crc kubenswrapper[4934]: E1002 09:49:09.858415 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:49:11.858409133 +0000 UTC m=+23.611050735 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.859796 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.859980 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.897954 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:09Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.912064 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.912121 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:09 crc kubenswrapper[4934]: E1002 09:49:09.912172 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.912138 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:09 crc kubenswrapper[4934]: E1002 09:49:09.912268 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:49:09 crc kubenswrapper[4934]: E1002 09:49:09.912346 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.917008 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:09Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.936274 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:09Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.948956 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:09Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.958508 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/2e2b3511-723e-4cb7-aaf6-1b077ae88458-hosts-file\") pod \"node-resolver-85mph\" (UID: \"2e2b3511-723e-4cb7-aaf6-1b077ae88458\") " pod="openshift-dns/node-resolver-85mph" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.958542 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf88n\" (UniqueName: \"kubernetes.io/projected/2e2b3511-723e-4cb7-aaf6-1b077ae88458-kube-api-access-lf88n\") pod \"node-resolver-85mph\" (UID: \"2e2b3511-723e-4cb7-aaf6-1b077ae88458\") " pod="openshift-dns/node-resolver-85mph" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.970688 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f950f5b95e378c306df980ebec95b90b817a0720a8582a2b12ef1533f71e43b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:02Z\\\",\\\"message\\\":\\\"W1002 09:48:51.991265 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 09:48:51.991593 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759398531 cert, and key in /tmp/serving-cert-3222084787/serving-signer.crt, /tmp/serving-cert-3222084787/serving-signer.key\\\\nI1002 09:48:52.252472 1 observer_polling.go:159] Starting file observer\\\\nW1002 09:48:52.256275 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 09:48:52.256431 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:48:52.257041 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3222084787/tls.crt::/tmp/serving-cert-3222084787/tls.key\\\\\\\"\\\\nF1002 09:49:02.761346 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:09Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.982165 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:09Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:09 crc kubenswrapper[4934]: I1002 09:49:09.994376 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:09Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.008514 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.024767 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.025097 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.027112 4934 scope.go:117] "RemoveContainer" containerID="a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97" Oct 02 09:49:10 crc kubenswrapper[4934]: E1002 09:49:10.027264 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.052434 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.059568 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/2e2b3511-723e-4cb7-aaf6-1b077ae88458-hosts-file\") pod \"node-resolver-85mph\" (UID: \"2e2b3511-723e-4cb7-aaf6-1b077ae88458\") " pod="openshift-dns/node-resolver-85mph" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.059658 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf88n\" (UniqueName: \"kubernetes.io/projected/2e2b3511-723e-4cb7-aaf6-1b077ae88458-kube-api-access-lf88n\") pod \"node-resolver-85mph\" (UID: \"2e2b3511-723e-4cb7-aaf6-1b077ae88458\") " pod="openshift-dns/node-resolver-85mph" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.059569 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/2e2b3511-723e-4cb7-aaf6-1b077ae88458-hosts-file\") pod \"node-resolver-85mph\" (UID: \"2e2b3511-723e-4cb7-aaf6-1b077ae88458\") " pod="openshift-dns/node-resolver-85mph" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.070456 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.079858 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf88n\" (UniqueName: \"kubernetes.io/projected/2e2b3511-723e-4cb7-aaf6-1b077ae88458-kube-api-access-lf88n\") pod \"node-resolver-85mph\" (UID: \"2e2b3511-723e-4cb7-aaf6-1b077ae88458\") " pod="openshift-dns/node-resolver-85mph" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.086764 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.102981 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.120376 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.133986 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.150436 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.162570 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-85mph" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.163216 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: W1002 09:49:10.174766 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e2b3511_723e_4cb7_aaf6_1b077ae88458.slice/crio-7685f7fa83af76d2d9a0899a9dcbe45b37e3a1eeffa16075ea101a4617d2a7bb WatchSource:0}: Error finding container 7685f7fa83af76d2d9a0899a9dcbe45b37e3a1eeffa16075ea101a4617d2a7bb: Status 404 returned error can't find the container with id 7685f7fa83af76d2d9a0899a9dcbe45b37e3a1eeffa16075ea101a4617d2a7bb Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.181390 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.361595 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.374809 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.380680 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.393986 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.404111 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.414932 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.427193 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.434388 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.445272 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.459434 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.475449 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.491274 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.514453 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.537250 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.556038 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.577615 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.605671 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.617192 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-djh5z"] Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.617616 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.621429 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.621558 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.621755 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.621849 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.633693 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.633937 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.634895 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-79fxg"] Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.635292 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-ds45z"] Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.635618 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.636153 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-th4v6"] Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.636928 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-ds45z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.637010 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.637043 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.641695 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.641791 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.642138 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.642157 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.642798 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.642838 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.642941 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.643004 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.643130 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.643716 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.643743 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.643807 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.643726 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.656765 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.673479 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.690420 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.713688 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.731134 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.743157 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.756266 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.764618 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-os-release\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.764664 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-host-var-lib-kubelet\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.764687 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-var-lib-openvswitch\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.764709 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1d4841b5-0469-461e-875c-25b9fe848141-ovnkube-config\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.764732 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-cnibin\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.764750 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-host-var-lib-cni-multus\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.764774 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-slash\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.764797 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-node-log\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.764820 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-kubelet\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.764839 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1d4841b5-0469-461e-875c-25b9fe848141-env-overrides\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.764860 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/73cb6891-fe6c-466e-a8b3-5497f28d2741-system-cni-dir\") pod \"multus-additional-cni-plugins-ds45z\" (UID: \"73cb6891-fe6c-466e-a8b3-5497f28d2741\") " pod="openshift-multus/multus-additional-cni-plugins-ds45z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.764880 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/73cb6891-fe6c-466e-a8b3-5497f28d2741-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ds45z\" (UID: \"73cb6891-fe6c-466e-a8b3-5497f28d2741\") " pod="openshift-multus/multus-additional-cni-plugins-ds45z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.764898 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxs8x\" (UniqueName: \"kubernetes.io/projected/73cb6891-fe6c-466e-a8b3-5497f28d2741-kube-api-access-cxs8x\") pod \"multus-additional-cni-plugins-ds45z\" (UID: \"73cb6891-fe6c-466e-a8b3-5497f28d2741\") " pod="openshift-multus/multus-additional-cni-plugins-ds45z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.764923 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-run-openvswitch\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.764941 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-cni-netd\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.764959 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-host-run-multus-certs\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.764976 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-etc-openvswitch\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.764995 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stfqs\" (UniqueName: \"kubernetes.io/projected/71db06ef-05b0-4f58-b251-b27117a8500a-kube-api-access-stfqs\") pod \"machine-config-daemon-djh5z\" (UID: \"71db06ef-05b0-4f58-b251-b27117a8500a\") " pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765009 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-system-cni-dir\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765024 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-run-ovn\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765038 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/73cb6891-fe6c-466e-a8b3-5497f28d2741-cni-binary-copy\") pod \"multus-additional-cni-plugins-ds45z\" (UID: \"73cb6891-fe6c-466e-a8b3-5497f28d2741\") " pod="openshift-multus/multus-additional-cni-plugins-ds45z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765052 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-etc-kubernetes\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765068 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-log-socket\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765132 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/71db06ef-05b0-4f58-b251-b27117a8500a-mcd-auth-proxy-config\") pod \"machine-config-daemon-djh5z\" (UID: \"71db06ef-05b0-4f58-b251-b27117a8500a\") " pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765175 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/73cb6891-fe6c-466e-a8b3-5497f28d2741-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ds45z\" (UID: \"73cb6891-fe6c-466e-a8b3-5497f28d2741\") " pod="openshift-multus/multus-additional-cni-plugins-ds45z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765199 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-host-run-netns\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765214 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1d4841b5-0469-461e-875c-25b9fe848141-ovnkube-script-lib\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765228 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/71db06ef-05b0-4f58-b251-b27117a8500a-rootfs\") pod \"machine-config-daemon-djh5z\" (UID: \"71db06ef-05b0-4f58-b251-b27117a8500a\") " pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765241 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/73cb6891-fe6c-466e-a8b3-5497f28d2741-os-release\") pod \"multus-additional-cni-plugins-ds45z\" (UID: \"73cb6891-fe6c-466e-a8b3-5497f28d2741\") " pod="openshift-multus/multus-additional-cni-plugins-ds45z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765257 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/73968247-67dd-48cc-88a1-64afac657412-multus-daemon-config\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765276 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-systemd-units\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765293 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-run-ovn-kubernetes\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765309 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765347 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-host-run-k8s-cni-cncf-io\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765369 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-multus-conf-dir\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765384 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/71db06ef-05b0-4f58-b251-b27117a8500a-proxy-tls\") pod \"machine-config-daemon-djh5z\" (UID: \"71db06ef-05b0-4f58-b251-b27117a8500a\") " pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765402 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-host-var-lib-cni-bin\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765440 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/73cb6891-fe6c-466e-a8b3-5497f28d2741-cnibin\") pod \"multus-additional-cni-plugins-ds45z\" (UID: \"73cb6891-fe6c-466e-a8b3-5497f28d2741\") " pod="openshift-multus/multus-additional-cni-plugins-ds45z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765496 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-multus-cni-dir\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765529 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mgrl\" (UniqueName: \"kubernetes.io/projected/1d4841b5-0469-461e-875c-25b9fe848141-kube-api-access-4mgrl\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765553 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-multus-socket-dir-parent\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765572 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-hostroot\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765612 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmk48\" (UniqueName: \"kubernetes.io/projected/73968247-67dd-48cc-88a1-64afac657412-kube-api-access-kmk48\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765633 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-cni-bin\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765652 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1d4841b5-0469-461e-875c-25b9fe848141-ovn-node-metrics-cert\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765671 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/73968247-67dd-48cc-88a1-64afac657412-cni-binary-copy\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765706 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-run-netns\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.765725 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-run-systemd\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.770070 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.783471 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.797471 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.811165 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.822303 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.840518 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.854484 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.867164 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-etc-kubernetes\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.867219 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-log-socket\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.867283 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-etc-kubernetes\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.867322 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-log-socket\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.867437 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/71db06ef-05b0-4f58-b251-b27117a8500a-mcd-auth-proxy-config\") pod \"machine-config-daemon-djh5z\" (UID: \"71db06ef-05b0-4f58-b251-b27117a8500a\") " pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.867472 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-host-run-netns\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.867502 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1d4841b5-0469-461e-875c-25b9fe848141-ovnkube-script-lib\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.867530 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/71db06ef-05b0-4f58-b251-b27117a8500a-rootfs\") pod \"machine-config-daemon-djh5z\" (UID: \"71db06ef-05b0-4f58-b251-b27117a8500a\") " pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.867551 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/73cb6891-fe6c-466e-a8b3-5497f28d2741-os-release\") pod \"multus-additional-cni-plugins-ds45z\" (UID: \"73cb6891-fe6c-466e-a8b3-5497f28d2741\") " pod="openshift-multus/multus-additional-cni-plugins-ds45z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.867593 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/73cb6891-fe6c-466e-a8b3-5497f28d2741-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ds45z\" (UID: \"73cb6891-fe6c-466e-a8b3-5497f28d2741\") " pod="openshift-multus/multus-additional-cni-plugins-ds45z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.867624 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/73968247-67dd-48cc-88a1-64afac657412-multus-daemon-config\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.867652 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-systemd-units\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.867676 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-run-ovn-kubernetes\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.867700 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.867749 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-host-run-k8s-cni-cncf-io\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.867773 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-multus-conf-dir\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.867800 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/71db06ef-05b0-4f58-b251-b27117a8500a-proxy-tls\") pod \"machine-config-daemon-djh5z\" (UID: \"71db06ef-05b0-4f58-b251-b27117a8500a\") " pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.867828 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-host-var-lib-cni-bin\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.867865 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/73cb6891-fe6c-466e-a8b3-5497f28d2741-cnibin\") pod \"multus-additional-cni-plugins-ds45z\" (UID: \"73cb6891-fe6c-466e-a8b3-5497f28d2741\") " pod="openshift-multus/multus-additional-cni-plugins-ds45z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.867933 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-multus-cni-dir\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.867958 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mgrl\" (UniqueName: \"kubernetes.io/projected/1d4841b5-0469-461e-875c-25b9fe848141-kube-api-access-4mgrl\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.867987 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-multus-socket-dir-parent\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868014 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-hostroot\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868040 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmk48\" (UniqueName: \"kubernetes.io/projected/73968247-67dd-48cc-88a1-64afac657412-kube-api-access-kmk48\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868069 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-cni-bin\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868091 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1d4841b5-0469-461e-875c-25b9fe848141-ovn-node-metrics-cert\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868123 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/73968247-67dd-48cc-88a1-64afac657412-cni-binary-copy\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868151 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-run-netns\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868179 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-run-systemd\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868235 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-os-release\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868258 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-host-var-lib-kubelet\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868265 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/71db06ef-05b0-4f58-b251-b27117a8500a-mcd-auth-proxy-config\") pod \"machine-config-daemon-djh5z\" (UID: \"71db06ef-05b0-4f58-b251-b27117a8500a\") " pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868287 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-var-lib-openvswitch\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868331 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-var-lib-openvswitch\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868333 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1d4841b5-0469-461e-875c-25b9fe848141-ovnkube-config\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868374 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-cnibin\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868406 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-host-var-lib-cni-multus\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868435 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-slash\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868456 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-node-log\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868482 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-kubelet\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868509 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1d4841b5-0469-461e-875c-25b9fe848141-env-overrides\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868537 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/73cb6891-fe6c-466e-a8b3-5497f28d2741-system-cni-dir\") pod \"multus-additional-cni-plugins-ds45z\" (UID: \"73cb6891-fe6c-466e-a8b3-5497f28d2741\") " pod="openshift-multus/multus-additional-cni-plugins-ds45z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868562 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/73cb6891-fe6c-466e-a8b3-5497f28d2741-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ds45z\" (UID: \"73cb6891-fe6c-466e-a8b3-5497f28d2741\") " pod="openshift-multus/multus-additional-cni-plugins-ds45z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868620 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxs8x\" (UniqueName: \"kubernetes.io/projected/73cb6891-fe6c-466e-a8b3-5497f28d2741-kube-api-access-cxs8x\") pod \"multus-additional-cni-plugins-ds45z\" (UID: \"73cb6891-fe6c-466e-a8b3-5497f28d2741\") " pod="openshift-multus/multus-additional-cni-plugins-ds45z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868651 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-cni-netd\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868683 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-host-run-multus-certs\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868713 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-etc-openvswitch\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868736 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-run-openvswitch\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868819 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stfqs\" (UniqueName: \"kubernetes.io/projected/71db06ef-05b0-4f58-b251-b27117a8500a-kube-api-access-stfqs\") pod \"machine-config-daemon-djh5z\" (UID: \"71db06ef-05b0-4f58-b251-b27117a8500a\") " pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868883 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-system-cni-dir\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868913 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-run-ovn\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868936 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/73cb6891-fe6c-466e-a8b3-5497f28d2741-cni-binary-copy\") pod \"multus-additional-cni-plugins-ds45z\" (UID: \"73cb6891-fe6c-466e-a8b3-5497f28d2741\") " pod="openshift-multus/multus-additional-cni-plugins-ds45z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.868977 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1d4841b5-0469-461e-875c-25b9fe848141-ovnkube-config\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.869074 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-host-run-k8s-cni-cncf-io\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.869096 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-host-run-multus-certs\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.869096 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-etc-openvswitch\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.869142 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-host-run-netns\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.869145 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-host-var-lib-kubelet\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.869180 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/73cb6891-fe6c-466e-a8b3-5497f28d2741-system-cni-dir\") pod \"multus-additional-cni-plugins-ds45z\" (UID: \"73cb6891-fe6c-466e-a8b3-5497f28d2741\") " pod="openshift-multus/multus-additional-cni-plugins-ds45z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.869264 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-multus-conf-dir\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.869393 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-multus-cni-dir\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.869424 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-run-netns\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.869478 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-run-systemd\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.869554 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1d4841b5-0469-461e-875c-25b9fe848141-env-overrides\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.869650 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-cnibin\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.869698 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-host-var-lib-cni-multus\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.869739 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-slash\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.869741 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-host-var-lib-cni-bin\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.869781 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-node-log\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.869787 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-kubelet\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.869817 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/73cb6891-fe6c-466e-a8b3-5497f28d2741-cnibin\") pod \"multus-additional-cni-plugins-ds45z\" (UID: \"73cb6891-fe6c-466e-a8b3-5497f28d2741\") " pod="openshift-multus/multus-additional-cni-plugins-ds45z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.869812 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/73cb6891-fe6c-466e-a8b3-5497f28d2741-cni-binary-copy\") pod \"multus-additional-cni-plugins-ds45z\" (UID: \"73cb6891-fe6c-466e-a8b3-5497f28d2741\") " pod="openshift-multus/multus-additional-cni-plugins-ds45z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.869887 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1d4841b5-0469-461e-875c-25b9fe848141-ovnkube-script-lib\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.869947 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/71db06ef-05b0-4f58-b251-b27117a8500a-rootfs\") pod \"machine-config-daemon-djh5z\" (UID: \"71db06ef-05b0-4f58-b251-b27117a8500a\") " pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.869961 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-run-openvswitch\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.870008 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/73cb6891-fe6c-466e-a8b3-5497f28d2741-os-release\") pod \"multus-additional-cni-plugins-ds45z\" (UID: \"73cb6891-fe6c-466e-a8b3-5497f28d2741\") " pod="openshift-multus/multus-additional-cni-plugins-ds45z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.870012 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-os-release\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.870016 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/73cb6891-fe6c-466e-a8b3-5497f28d2741-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ds45z\" (UID: \"73cb6891-fe6c-466e-a8b3-5497f28d2741\") " pod="openshift-multus/multus-additional-cni-plugins-ds45z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.870040 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-systemd-units\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.870073 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-run-ovn-kubernetes\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.870128 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-hostroot\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.870121 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.870201 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-multus-socket-dir-parent\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.870221 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/73968247-67dd-48cc-88a1-64afac657412-system-cni-dir\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.870242 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/73968247-67dd-48cc-88a1-64afac657412-cni-binary-copy\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.870250 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-cni-bin\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.870281 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-run-ovn\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.870280 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-cni-netd\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.870680 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/73968247-67dd-48cc-88a1-64afac657412-multus-daemon-config\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.871033 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.873741 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/71db06ef-05b0-4f58-b251-b27117a8500a-proxy-tls\") pod \"machine-config-daemon-djh5z\" (UID: \"71db06ef-05b0-4f58-b251-b27117a8500a\") " pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.874787 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1d4841b5-0469-461e-875c-25b9fe848141-ovn-node-metrics-cert\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.880769 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.888008 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/73cb6891-fe6c-466e-a8b3-5497f28d2741-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ds45z\" (UID: \"73cb6891-fe6c-466e-a8b3-5497f28d2741\") " pod="openshift-multus/multus-additional-cni-plugins-ds45z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.888116 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmk48\" (UniqueName: \"kubernetes.io/projected/73968247-67dd-48cc-88a1-64afac657412-kube-api-access-kmk48\") pod \"multus-79fxg\" (UID: \"73968247-67dd-48cc-88a1-64afac657412\") " pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.888482 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stfqs\" (UniqueName: \"kubernetes.io/projected/71db06ef-05b0-4f58-b251-b27117a8500a-kube-api-access-stfqs\") pod \"machine-config-daemon-djh5z\" (UID: \"71db06ef-05b0-4f58-b251-b27117a8500a\") " pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.888897 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxs8x\" (UniqueName: \"kubernetes.io/projected/73cb6891-fe6c-466e-a8b3-5497f28d2741-kube-api-access-cxs8x\") pod \"multus-additional-cni-plugins-ds45z\" (UID: \"73cb6891-fe6c-466e-a8b3-5497f28d2741\") " pod="openshift-multus/multus-additional-cni-plugins-ds45z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.890378 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mgrl\" (UniqueName: \"kubernetes.io/projected/1d4841b5-0469-461e-875c-25b9fe848141-kube-api-access-4mgrl\") pod \"ovnkube-node-th4v6\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.905925 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.925835 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.948132 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.955699 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.964925 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-79fxg" Oct 02 09:49:10 crc kubenswrapper[4934]: I1002 09:49:10.971171 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-ds45z" Oct 02 09:49:10 crc kubenswrapper[4934]: W1002 09:49:10.974889 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71db06ef_05b0_4f58_b251_b27117a8500a.slice/crio-482a23930b998efec0a8fc9c11ca19c3a5a890e939e29d0e4bc67886905d10d1 WatchSource:0}: Error finding container 482a23930b998efec0a8fc9c11ca19c3a5a890e939e29d0e4bc67886905d10d1: Status 404 returned error can't find the container with id 482a23930b998efec0a8fc9c11ca19c3a5a890e939e29d0e4bc67886905d10d1 Oct 02 09:49:10 crc kubenswrapper[4934]: W1002 09:49:10.985344 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73cb6891_fe6c_466e_a8b3_5497f28d2741.slice/crio-728b11f2f32bd59a96990b52074ac3790170dec847df6a49b608b0c95ab206e0 WatchSource:0}: Error finding container 728b11f2f32bd59a96990b52074ac3790170dec847df6a49b608b0c95ab206e0: Status 404 returned error can't find the container with id 728b11f2f32bd59a96990b52074ac3790170dec847df6a49b608b0c95ab206e0 Oct 02 09:49:10 crc kubenswrapper[4934]: W1002 09:49:10.986558 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d4841b5_0469_461e_875c_25b9fe848141.slice/crio-bcb32bcd13898c8fdf7e282c3b8794230b382612943869b10216ed0361924b16 WatchSource:0}: Error finding container bcb32bcd13898c8fdf7e282c3b8794230b382612943869b10216ed0361924b16: Status 404 returned error can't find the container with id bcb32bcd13898c8fdf7e282c3b8794230b382612943869b10216ed0361924b16 Oct 02 09:49:10 crc kubenswrapper[4934]: W1002 09:49:10.989763 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73968247_67dd_48cc_88a1_64afac657412.slice/crio-6efffd378da46cbe7a201e61d08ec297fa0ec96384490e6a7b4441795c1ee2ab WatchSource:0}: Error finding container 6efffd378da46cbe7a201e61d08ec297fa0ec96384490e6a7b4441795c1ee2ab: Status 404 returned error can't find the container with id 6efffd378da46cbe7a201e61d08ec297fa0ec96384490e6a7b4441795c1ee2ab Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.030808 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" event={"ID":"1d4841b5-0469-461e-875c-25b9fe848141","Type":"ContainerStarted","Data":"bcb32bcd13898c8fdf7e282c3b8794230b382612943869b10216ed0361924b16"} Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.032909 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" event={"ID":"73cb6891-fe6c-466e-a8b3-5497f28d2741","Type":"ContainerStarted","Data":"728b11f2f32bd59a96990b52074ac3790170dec847df6a49b608b0c95ab206e0"} Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.033865 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerStarted","Data":"482a23930b998efec0a8fc9c11ca19c3a5a890e939e29d0e4bc67886905d10d1"} Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.035093 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2"} Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.036067 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-79fxg" event={"ID":"73968247-67dd-48cc-88a1-64afac657412","Type":"ContainerStarted","Data":"6efffd378da46cbe7a201e61d08ec297fa0ec96384490e6a7b4441795c1ee2ab"} Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.038856 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-85mph" event={"ID":"2e2b3511-723e-4cb7-aaf6-1b077ae88458","Type":"ContainerStarted","Data":"51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e"} Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.039307 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-85mph" event={"ID":"2e2b3511-723e-4cb7-aaf6-1b077ae88458","Type":"ContainerStarted","Data":"7685f7fa83af76d2d9a0899a9dcbe45b37e3a1eeffa16075ea101a4617d2a7bb"} Oct 02 09:49:11 crc kubenswrapper[4934]: E1002 09:49:11.047963 4934 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.049451 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.061487 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.075876 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.087397 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.101118 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.113997 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.129325 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.141796 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.161775 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.175938 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.189238 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.200292 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.248342 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.285540 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.320806 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.360215 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.402402 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.443803 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.482701 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.526546 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.561805 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.604124 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.640143 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.687121 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.725917 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.762302 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.779092 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:11 crc kubenswrapper[4934]: E1002 09:49:11.779231 4934 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:49:11 crc kubenswrapper[4934]: E1002 09:49:11.779284 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:49:15.779271272 +0000 UTC m=+27.531912794 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.799883 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.844518 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.880002 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.880112 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.880154 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.880182 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:11 crc kubenswrapper[4934]: E1002 09:49:11.880210 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:49:15.880179571 +0000 UTC m=+27.632821093 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:49:11 crc kubenswrapper[4934]: E1002 09:49:11.880276 4934 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:49:11 crc kubenswrapper[4934]: E1002 09:49:11.880345 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:49:15.880327074 +0000 UTC m=+27.632968656 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:49:11 crc kubenswrapper[4934]: E1002 09:49:11.880354 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:49:11 crc kubenswrapper[4934]: E1002 09:49:11.880369 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:49:11 crc kubenswrapper[4934]: E1002 09:49:11.880381 4934 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:49:11 crc kubenswrapper[4934]: E1002 09:49:11.880375 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:49:11 crc kubenswrapper[4934]: E1002 09:49:11.880417 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 09:49:15.880409086 +0000 UTC m=+27.633050608 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:49:11 crc kubenswrapper[4934]: E1002 09:49:11.880421 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:49:11 crc kubenswrapper[4934]: E1002 09:49:11.880437 4934 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:49:11 crc kubenswrapper[4934]: E1002 09:49:11.880517 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 09:49:15.880497058 +0000 UTC m=+27.633138640 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.913037 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.913152 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:11 crc kubenswrapper[4934]: E1002 09:49:11.913166 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:49:11 crc kubenswrapper[4934]: I1002 09:49:11.913037 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:11 crc kubenswrapper[4934]: E1002 09:49:11.913299 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:49:11 crc kubenswrapper[4934]: E1002 09:49:11.913487 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.043028 4934 generic.go:334] "Generic (PLEG): container finished" podID="73cb6891-fe6c-466e-a8b3-5497f28d2741" containerID="880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8" exitCode=0 Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.043080 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" event={"ID":"73cb6891-fe6c-466e-a8b3-5497f28d2741","Type":"ContainerDied","Data":"880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8"} Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.049651 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerStarted","Data":"d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a"} Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.049717 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerStarted","Data":"42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af"} Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.052040 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-79fxg" event={"ID":"73968247-67dd-48cc-88a1-64afac657412","Type":"ContainerStarted","Data":"08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65"} Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.052980 4934 generic.go:334] "Generic (PLEG): container finished" podID="1d4841b5-0469-461e-875c-25b9fe848141" containerID="46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478" exitCode=0 Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.053825 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" event={"ID":"1d4841b5-0469-461e-875c-25b9fe848141","Type":"ContainerDied","Data":"46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478"} Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.066402 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.081931 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.097308 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.110065 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.123750 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.147165 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.161258 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.175106 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.200560 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.241769 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.280114 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.318982 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.361617 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.407648 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.442628 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.477187 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-tctph"] Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.477788 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-tctph" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.487938 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.492329 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.512941 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.533136 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.552654 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.587541 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/13155c94-2cc0-483d-afa0-68f2415404b0-serviceca\") pod \"node-ca-tctph\" (UID: \"13155c94-2cc0-483d-afa0-68f2415404b0\") " pod="openshift-image-registry/node-ca-tctph" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.588105 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pd7nm\" (UniqueName: \"kubernetes.io/projected/13155c94-2cc0-483d-afa0-68f2415404b0-kube-api-access-pd7nm\") pod \"node-ca-tctph\" (UID: \"13155c94-2cc0-483d-afa0-68f2415404b0\") " pod="openshift-image-registry/node-ca-tctph" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.588213 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/13155c94-2cc0-483d-afa0-68f2415404b0-host\") pod \"node-ca-tctph\" (UID: \"13155c94-2cc0-483d-afa0-68f2415404b0\") " pod="openshift-image-registry/node-ca-tctph" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.603692 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.651921 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.684667 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.689381 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/13155c94-2cc0-483d-afa0-68f2415404b0-serviceca\") pod \"node-ca-tctph\" (UID: \"13155c94-2cc0-483d-afa0-68f2415404b0\") " pod="openshift-image-registry/node-ca-tctph" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.689449 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pd7nm\" (UniqueName: \"kubernetes.io/projected/13155c94-2cc0-483d-afa0-68f2415404b0-kube-api-access-pd7nm\") pod \"node-ca-tctph\" (UID: \"13155c94-2cc0-483d-afa0-68f2415404b0\") " pod="openshift-image-registry/node-ca-tctph" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.689488 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/13155c94-2cc0-483d-afa0-68f2415404b0-host\") pod \"node-ca-tctph\" (UID: \"13155c94-2cc0-483d-afa0-68f2415404b0\") " pod="openshift-image-registry/node-ca-tctph" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.689559 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/13155c94-2cc0-483d-afa0-68f2415404b0-host\") pod \"node-ca-tctph\" (UID: \"13155c94-2cc0-483d-afa0-68f2415404b0\") " pod="openshift-image-registry/node-ca-tctph" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.691026 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/13155c94-2cc0-483d-afa0-68f2415404b0-serviceca\") pod \"node-ca-tctph\" (UID: \"13155c94-2cc0-483d-afa0-68f2415404b0\") " pod="openshift-image-registry/node-ca-tctph" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.732475 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pd7nm\" (UniqueName: \"kubernetes.io/projected/13155c94-2cc0-483d-afa0-68f2415404b0-kube-api-access-pd7nm\") pod \"node-ca-tctph\" (UID: \"13155c94-2cc0-483d-afa0-68f2415404b0\") " pod="openshift-image-registry/node-ca-tctph" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.742728 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.781872 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.823780 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.863744 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.886851 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-tctph" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.910884 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.944245 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:12 crc kubenswrapper[4934]: I1002 09:49:12.990065 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:13 crc kubenswrapper[4934]: I1002 09:49:13.028366 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:13 crc kubenswrapper[4934]: I1002 09:49:13.060620 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" event={"ID":"73cb6891-fe6c-466e-a8b3-5497f28d2741","Type":"ContainerStarted","Data":"0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a"} Oct 02 09:49:13 crc kubenswrapper[4934]: I1002 09:49:13.061900 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-tctph" event={"ID":"13155c94-2cc0-483d-afa0-68f2415404b0","Type":"ContainerStarted","Data":"06b093dd6ce896abc307ca9f3a2055b9ada57ab236e3212eca70f0fffb9996fe"} Oct 02 09:49:13 crc kubenswrapper[4934]: I1002 09:49:13.064347 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:13 crc kubenswrapper[4934]: I1002 09:49:13.067460 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" event={"ID":"1d4841b5-0469-461e-875c-25b9fe848141","Type":"ContainerStarted","Data":"9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea"} Oct 02 09:49:13 crc kubenswrapper[4934]: I1002 09:49:13.067509 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" event={"ID":"1d4841b5-0469-461e-875c-25b9fe848141","Type":"ContainerStarted","Data":"c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8"} Oct 02 09:49:13 crc kubenswrapper[4934]: I1002 09:49:13.067522 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" event={"ID":"1d4841b5-0469-461e-875c-25b9fe848141","Type":"ContainerStarted","Data":"389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9"} Oct 02 09:49:13 crc kubenswrapper[4934]: I1002 09:49:13.067536 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" event={"ID":"1d4841b5-0469-461e-875c-25b9fe848141","Type":"ContainerStarted","Data":"e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b"} Oct 02 09:49:13 crc kubenswrapper[4934]: I1002 09:49:13.067546 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" event={"ID":"1d4841b5-0469-461e-875c-25b9fe848141","Type":"ContainerStarted","Data":"e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019"} Oct 02 09:49:13 crc kubenswrapper[4934]: I1002 09:49:13.067556 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" event={"ID":"1d4841b5-0469-461e-875c-25b9fe848141","Type":"ContainerStarted","Data":"4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb"} Oct 02 09:49:13 crc kubenswrapper[4934]: I1002 09:49:13.104944 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:13 crc kubenswrapper[4934]: I1002 09:49:13.141356 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:13 crc kubenswrapper[4934]: I1002 09:49:13.180385 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:13 crc kubenswrapper[4934]: I1002 09:49:13.221329 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:13 crc kubenswrapper[4934]: I1002 09:49:13.260817 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:13 crc kubenswrapper[4934]: I1002 09:49:13.300858 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:13 crc kubenswrapper[4934]: I1002 09:49:13.342779 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:13 crc kubenswrapper[4934]: I1002 09:49:13.382616 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:13 crc kubenswrapper[4934]: I1002 09:49:13.422572 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:13 crc kubenswrapper[4934]: I1002 09:49:13.467961 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:13 crc kubenswrapper[4934]: I1002 09:49:13.500548 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:13 crc kubenswrapper[4934]: I1002 09:49:13.543279 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:13 crc kubenswrapper[4934]: I1002 09:49:13.591868 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:13 crc kubenswrapper[4934]: I1002 09:49:13.628204 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:13 crc kubenswrapper[4934]: I1002 09:49:13.662270 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:13 crc kubenswrapper[4934]: I1002 09:49:13.912450 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:13 crc kubenswrapper[4934]: I1002 09:49:13.912546 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:13 crc kubenswrapper[4934]: I1002 09:49:13.912681 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:13 crc kubenswrapper[4934]: E1002 09:49:13.912678 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:49:13 crc kubenswrapper[4934]: E1002 09:49:13.912825 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:49:13 crc kubenswrapper[4934]: E1002 09:49:13.912943 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.072646 4934 generic.go:334] "Generic (PLEG): container finished" podID="73cb6891-fe6c-466e-a8b3-5497f28d2741" containerID="0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a" exitCode=0 Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.072719 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" event={"ID":"73cb6891-fe6c-466e-a8b3-5497f28d2741","Type":"ContainerDied","Data":"0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a"} Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.074817 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-tctph" event={"ID":"13155c94-2cc0-483d-afa0-68f2415404b0","Type":"ContainerStarted","Data":"7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52"} Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.086607 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.102156 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.115516 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.129566 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.140766 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.156311 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.168538 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.180431 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.191670 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.212097 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.224902 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.238821 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.249569 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.267158 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.281343 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.307203 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.340782 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.381202 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.420823 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.460234 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.500960 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.529262 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.530892 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.530935 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.530947 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.531065 4934 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.541519 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.593233 4934 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.593477 4934 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.594597 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.594639 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.594652 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.594670 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.594683 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:14Z","lastTransitionTime":"2025-10-02T09:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:14 crc kubenswrapper[4934]: E1002 09:49:14.607500 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.611374 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.611412 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.611423 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.611441 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.611452 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:14Z","lastTransitionTime":"2025-10-02T09:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.619758 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: E1002 09:49:14.621709 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.625825 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.625883 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.625898 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.625918 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.625940 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:14Z","lastTransitionTime":"2025-10-02T09:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:14 crc kubenswrapper[4934]: E1002 09:49:14.638258 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.641706 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.641743 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.641753 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.641768 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.641778 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:14Z","lastTransitionTime":"2025-10-02T09:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:14 crc kubenswrapper[4934]: E1002 09:49:14.651899 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.655210 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.655257 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.655270 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.655288 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.655299 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:14Z","lastTransitionTime":"2025-10-02T09:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.661111 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: E1002 09:49:14.668863 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: E1002 09:49:14.669025 4934 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.670731 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.670777 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.670793 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.670813 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.670829 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:14Z","lastTransitionTime":"2025-10-02T09:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.700914 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.742476 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.773344 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.773389 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.773401 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.773418 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.773430 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:14Z","lastTransitionTime":"2025-10-02T09:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.780190 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.819401 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.870539 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.876926 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.876972 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.876982 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.877000 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.877012 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:14Z","lastTransitionTime":"2025-10-02T09:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.924712 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.979644 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.979692 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.979703 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.979721 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:14 crc kubenswrapper[4934]: I1002 09:49:14.979732 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:14Z","lastTransitionTime":"2025-10-02T09:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.080483 4934 generic.go:334] "Generic (PLEG): container finished" podID="73cb6891-fe6c-466e-a8b3-5497f28d2741" containerID="48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7" exitCode=0 Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.081019 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" event={"ID":"73cb6891-fe6c-466e-a8b3-5497f28d2741","Type":"ContainerDied","Data":"48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7"} Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.081759 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.081785 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.081796 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.081812 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.081824 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:15Z","lastTransitionTime":"2025-10-02T09:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.099932 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.112735 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.123886 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.138297 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.155136 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.175534 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.185017 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.185054 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.185065 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.185083 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.185094 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:15Z","lastTransitionTime":"2025-10-02T09:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.191185 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.223194 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.262308 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.287603 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.287643 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.287652 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.287666 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.287676 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:15Z","lastTransitionTime":"2025-10-02T09:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.302214 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.340971 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.381107 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.389892 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.389945 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.389955 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.389970 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.389981 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:15Z","lastTransitionTime":"2025-10-02T09:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.420571 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.462524 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.491768 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.491849 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.491866 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.491888 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.491904 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:15Z","lastTransitionTime":"2025-10-02T09:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.499783 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.595172 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.595220 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.595236 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.595260 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.595276 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:15Z","lastTransitionTime":"2025-10-02T09:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.698945 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.699016 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.699036 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.699065 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.699085 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:15Z","lastTransitionTime":"2025-10-02T09:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.801522 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.801622 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.801657 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.801688 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.801712 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:15Z","lastTransitionTime":"2025-10-02T09:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.819037 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:15 crc kubenswrapper[4934]: E1002 09:49:15.819176 4934 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:49:15 crc kubenswrapper[4934]: E1002 09:49:15.819231 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:49:23.819218845 +0000 UTC m=+35.571860367 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.904432 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.904473 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.904483 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.904498 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.904508 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:15Z","lastTransitionTime":"2025-10-02T09:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.913125 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.913153 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.913229 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:15 crc kubenswrapper[4934]: E1002 09:49:15.913264 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:49:15 crc kubenswrapper[4934]: E1002 09:49:15.913430 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:49:15 crc kubenswrapper[4934]: E1002 09:49:15.913689 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.919824 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.919988 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:15 crc kubenswrapper[4934]: E1002 09:49:15.920015 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:49:23.919991081 +0000 UTC m=+35.672632633 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.920056 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:15 crc kubenswrapper[4934]: I1002 09:49:15.920149 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:15 crc kubenswrapper[4934]: E1002 09:49:15.920150 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:49:15 crc kubenswrapper[4934]: E1002 09:49:15.920223 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:49:15 crc kubenswrapper[4934]: E1002 09:49:15.920248 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:49:15 crc kubenswrapper[4934]: E1002 09:49:15.920257 4934 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:49:15 crc kubenswrapper[4934]: E1002 09:49:15.920344 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:49:23.920319168 +0000 UTC m=+35.672960730 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:49:15 crc kubenswrapper[4934]: E1002 09:49:15.920266 4934 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:49:15 crc kubenswrapper[4934]: E1002 09:49:15.920424 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 09:49:23.92041077 +0000 UTC m=+35.673052332 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:49:15 crc kubenswrapper[4934]: E1002 09:49:15.920223 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:49:15 crc kubenswrapper[4934]: E1002 09:49:15.920460 4934 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:49:15 crc kubenswrapper[4934]: E1002 09:49:15.920511 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 09:49:23.920496802 +0000 UTC m=+35.673138364 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.007322 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.007391 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.007413 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.007441 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.007462 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:16Z","lastTransitionTime":"2025-10-02T09:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.016938 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.018207 4934 scope.go:117] "RemoveContainer" containerID="a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97" Oct 02 09:49:16 crc kubenswrapper[4934]: E1002 09:49:16.018532 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.089074 4934 generic.go:334] "Generic (PLEG): container finished" podID="73cb6891-fe6c-466e-a8b3-5497f28d2741" containerID="efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495" exitCode=0 Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.089191 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" event={"ID":"73cb6891-fe6c-466e-a8b3-5497f28d2741","Type":"ContainerDied","Data":"efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495"} Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.094909 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" event={"ID":"1d4841b5-0469-461e-875c-25b9fe848141","Type":"ContainerStarted","Data":"2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c"} Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.110045 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.110117 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.110140 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.110170 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.110195 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:16Z","lastTransitionTime":"2025-10-02T09:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.113850 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.133984 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.147061 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.162430 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.175064 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.185947 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.211298 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.213541 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.213639 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.213657 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.213680 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.213701 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:16Z","lastTransitionTime":"2025-10-02T09:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.229469 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.242282 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.252189 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.270439 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.283434 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.296060 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.307266 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.317104 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.317137 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.317145 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.317158 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.317167 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:16Z","lastTransitionTime":"2025-10-02T09:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.318117 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.420044 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.420089 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.420101 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.420119 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.420132 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:16Z","lastTransitionTime":"2025-10-02T09:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.522668 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.522952 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.523044 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.523124 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.523246 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:16Z","lastTransitionTime":"2025-10-02T09:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.625836 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.626046 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.626113 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.626215 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.626272 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:16Z","lastTransitionTime":"2025-10-02T09:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.728823 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.728863 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.728872 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.728889 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.728901 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:16Z","lastTransitionTime":"2025-10-02T09:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.831927 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.832253 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.832455 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.832722 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.832975 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:16Z","lastTransitionTime":"2025-10-02T09:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.935736 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.935790 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.935826 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.935851 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:16 crc kubenswrapper[4934]: I1002 09:49:16.935869 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:16Z","lastTransitionTime":"2025-10-02T09:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.038213 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.038247 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.038256 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.038270 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.038280 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:17Z","lastTransitionTime":"2025-10-02T09:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.100370 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" event={"ID":"73cb6891-fe6c-466e-a8b3-5497f28d2741","Type":"ContainerStarted","Data":"4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897"} Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.121364 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:17Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.136565 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:17Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.140386 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.140417 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.140427 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.140443 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.140455 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:17Z","lastTransitionTime":"2025-10-02T09:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.147555 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:17Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.159894 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:17Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.181558 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:17Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.200023 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:17Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.216212 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:17Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.232186 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:17Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.242933 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.242983 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.242994 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.243010 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.243021 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:17Z","lastTransitionTime":"2025-10-02T09:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.247684 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:17Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.258990 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:17Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.271384 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:17Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.286486 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:17Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.298753 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:17Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.311098 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:17Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.321287 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:17Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.345411 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.345456 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.345467 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.345483 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.345493 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:17Z","lastTransitionTime":"2025-10-02T09:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.447716 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.447759 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.447775 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.447794 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.447805 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:17Z","lastTransitionTime":"2025-10-02T09:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.550747 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.550800 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.550813 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.550835 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.550847 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:17Z","lastTransitionTime":"2025-10-02T09:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.653392 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.653525 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.653651 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.653870 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.653979 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:17Z","lastTransitionTime":"2025-10-02T09:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.756936 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.757228 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.757427 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.757696 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.757907 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:17Z","lastTransitionTime":"2025-10-02T09:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.861369 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.861420 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.861438 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.861462 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.861482 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:17Z","lastTransitionTime":"2025-10-02T09:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.912455 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.912510 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.912462 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:17 crc kubenswrapper[4934]: E1002 09:49:17.912689 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:49:17 crc kubenswrapper[4934]: E1002 09:49:17.912802 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:49:17 crc kubenswrapper[4934]: E1002 09:49:17.912921 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.965142 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.965198 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.965215 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.965239 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:17 crc kubenswrapper[4934]: I1002 09:49:17.965260 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:17Z","lastTransitionTime":"2025-10-02T09:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.068916 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.069327 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.069339 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.069358 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.069369 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:18Z","lastTransitionTime":"2025-10-02T09:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.107443 4934 generic.go:334] "Generic (PLEG): container finished" podID="73cb6891-fe6c-466e-a8b3-5497f28d2741" containerID="4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897" exitCode=0 Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.107536 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" event={"ID":"73cb6891-fe6c-466e-a8b3-5497f28d2741","Type":"ContainerDied","Data":"4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897"} Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.113860 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" event={"ID":"1d4841b5-0469-461e-875c-25b9fe848141","Type":"ContainerStarted","Data":"cc415592d60854b2615100fa4a982146a0b2b770d6fe9e4a61065d13cddb75dc"} Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.114342 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.114399 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.140769 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.149028 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.149815 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.160654 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.172545 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.172618 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.172629 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.172666 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.172677 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:18Z","lastTransitionTime":"2025-10-02T09:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.175319 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.188950 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.203246 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.215945 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.229200 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.243942 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.254345 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.263866 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.274704 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.274736 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.274745 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.274760 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.274771 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:18Z","lastTransitionTime":"2025-10-02T09:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.275933 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.287124 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.296879 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.308015 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.325893 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.338719 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.351314 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.363405 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.373126 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.377534 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.377592 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.377602 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.377616 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.377627 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:18Z","lastTransitionTime":"2025-10-02T09:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.386084 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.397556 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.408355 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.418955 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.432657 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.445837 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.456727 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.469143 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.480335 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.480380 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.480391 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.480407 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.480417 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:18Z","lastTransitionTime":"2025-10-02T09:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.487562 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc415592d60854b2615100fa4a982146a0b2b770d6fe9e4a61065d13cddb75dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.511131 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.525624 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.582781 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.582827 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.582842 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.582861 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.582873 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:18Z","lastTransitionTime":"2025-10-02T09:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.685139 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.685215 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.685230 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.685258 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.685271 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:18Z","lastTransitionTime":"2025-10-02T09:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.787480 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.787533 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.787545 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.787562 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.787589 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:18Z","lastTransitionTime":"2025-10-02T09:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.893490 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.893537 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.893548 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.893565 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.893603 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:18Z","lastTransitionTime":"2025-10-02T09:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.926678 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.940356 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.951383 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.962617 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.977116 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.988888 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.996284 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.996570 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.996682 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.996770 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.996839 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:18Z","lastTransitionTime":"2025-10-02T09:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:18 crc kubenswrapper[4934]: I1002 09:49:18.997902 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.011004 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.021817 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.034188 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.051681 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc415592d60854b2615100fa4a982146a0b2b770d6fe9e4a61065d13cddb75dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.063887 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.076319 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.091309 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.099243 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.099294 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.099314 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.099334 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.099345 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:19Z","lastTransitionTime":"2025-10-02T09:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.112441 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.119538 4934 generic.go:334] "Generic (PLEG): container finished" podID="73cb6891-fe6c-466e-a8b3-5497f28d2741" containerID="76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1" exitCode=0 Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.119610 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" event={"ID":"73cb6891-fe6c-466e-a8b3-5497f28d2741","Type":"ContainerDied","Data":"76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1"} Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.119711 4934 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.137874 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.151889 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.161217 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.171561 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.190535 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc415592d60854b2615100fa4a982146a0b2b770d6fe9e4a61065d13cddb75dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.201741 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.201773 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.201780 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.201793 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.201802 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:19Z","lastTransitionTime":"2025-10-02T09:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.209898 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.222996 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.234376 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.245725 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.257159 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.268789 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.280793 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.293168 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.304221 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.304262 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.304279 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.304293 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.304303 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:19Z","lastTransitionTime":"2025-10-02T09:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.304123 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.314852 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.406263 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.406312 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.406324 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.406343 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.406353 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:19Z","lastTransitionTime":"2025-10-02T09:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.509793 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.510141 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.510156 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.510198 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.510210 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:19Z","lastTransitionTime":"2025-10-02T09:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.612534 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.612569 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.612598 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.612614 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.612624 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:19Z","lastTransitionTime":"2025-10-02T09:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.715302 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.715332 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.715343 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.715356 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.715366 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:19Z","lastTransitionTime":"2025-10-02T09:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.818186 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.818245 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.818257 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.818274 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.818609 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:19Z","lastTransitionTime":"2025-10-02T09:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.912652 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.912704 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.912666 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:19 crc kubenswrapper[4934]: E1002 09:49:19.912797 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:49:19 crc kubenswrapper[4934]: E1002 09:49:19.912835 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:49:19 crc kubenswrapper[4934]: E1002 09:49:19.912910 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.921662 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.921710 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.921723 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.921739 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:19 crc kubenswrapper[4934]: I1002 09:49:19.921751 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:19Z","lastTransitionTime":"2025-10-02T09:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.024109 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.024164 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.024174 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.024189 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.024200 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:20Z","lastTransitionTime":"2025-10-02T09:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.125569 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.125614 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.125623 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.125637 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.125645 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:20Z","lastTransitionTime":"2025-10-02T09:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.126370 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" event={"ID":"73cb6891-fe6c-466e-a8b3-5497f28d2741","Type":"ContainerStarted","Data":"851e2f96cda63332af97a90a26088ecdd7fcf2d0498b53e5d542aa532b048e20"} Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.126407 4934 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.143785 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.156937 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851e2f96cda63332af97a90a26088ecdd7fcf2d0498b53e5d542aa532b048e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.173365 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.186958 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.198541 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.213616 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.223768 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.227784 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.227877 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.227891 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.227913 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.227926 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:20Z","lastTransitionTime":"2025-10-02T09:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.235092 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.247065 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.257325 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.269404 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.280167 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.288501 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.297871 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.313500 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc415592d60854b2615100fa4a982146a0b2b770d6fe9e4a61065d13cddb75dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:20Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.330592 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.330624 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.330634 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.330650 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.330661 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:20Z","lastTransitionTime":"2025-10-02T09:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.432886 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.432917 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.432927 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.432942 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.432952 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:20Z","lastTransitionTime":"2025-10-02T09:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.535023 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.535058 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.535070 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.535085 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.535095 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:20Z","lastTransitionTime":"2025-10-02T09:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.637230 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.637266 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.637276 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.637290 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.637301 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:20Z","lastTransitionTime":"2025-10-02T09:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.739181 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.739256 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.739268 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.739288 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.739299 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:20Z","lastTransitionTime":"2025-10-02T09:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.842674 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.842721 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.842737 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.842752 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.842762 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:20Z","lastTransitionTime":"2025-10-02T09:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.944846 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.944911 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.944934 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.944964 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:20 crc kubenswrapper[4934]: I1002 09:49:20.944988 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:20Z","lastTransitionTime":"2025-10-02T09:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.047796 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.047833 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.047844 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.047866 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.047879 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:21Z","lastTransitionTime":"2025-10-02T09:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.130713 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-th4v6_1d4841b5-0469-461e-875c-25b9fe848141/ovnkube-controller/0.log" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.133744 4934 generic.go:334] "Generic (PLEG): container finished" podID="1d4841b5-0469-461e-875c-25b9fe848141" containerID="cc415592d60854b2615100fa4a982146a0b2b770d6fe9e4a61065d13cddb75dc" exitCode=1 Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.133759 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" event={"ID":"1d4841b5-0469-461e-875c-25b9fe848141","Type":"ContainerDied","Data":"cc415592d60854b2615100fa4a982146a0b2b770d6fe9e4a61065d13cddb75dc"} Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.134784 4934 scope.go:117] "RemoveContainer" containerID="cc415592d60854b2615100fa4a982146a0b2b770d6fe9e4a61065d13cddb75dc" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.150621 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.150655 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.150666 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.150681 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.150692 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:21Z","lastTransitionTime":"2025-10-02T09:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.158326 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc415592d60854b2615100fa4a982146a0b2b770d6fe9e4a61065d13cddb75dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc415592d60854b2615100fa4a982146a0b2b770d6fe9e4a61065d13cddb75dc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"message\\\":\\\"kAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1002 09:49:20.272194 6200 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:20.272253 6200 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:49:20.272486 6200 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:49:20.272672 6200 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:20.272711 6200 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:49:20.272883 6200 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:49:20.273167 6200 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.171729 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.184804 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.195939 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.207184 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.229719 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.243891 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851e2f96cda63332af97a90a26088ecdd7fcf2d0498b53e5d542aa532b048e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.253131 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.253162 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.253172 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.253186 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.253194 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:21Z","lastTransitionTime":"2025-10-02T09:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.266169 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.278194 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.291482 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.303668 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.313699 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.325273 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.336360 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.348447 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.354895 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.354927 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.354936 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.354949 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.354959 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:21Z","lastTransitionTime":"2025-10-02T09:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.457603 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.457641 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.457650 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.457664 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.457673 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:21Z","lastTransitionTime":"2025-10-02T09:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.560297 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.560340 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.560352 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.560372 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.560385 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:21Z","lastTransitionTime":"2025-10-02T09:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.662620 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.662663 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.662673 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.662689 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.662698 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:21Z","lastTransitionTime":"2025-10-02T09:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.772353 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.772390 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.772400 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.772415 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.772426 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:21Z","lastTransitionTime":"2025-10-02T09:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.874824 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.874866 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.874878 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.874901 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.874912 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:21Z","lastTransitionTime":"2025-10-02T09:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.913021 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.913075 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.913028 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:21 crc kubenswrapper[4934]: E1002 09:49:21.913241 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:49:21 crc kubenswrapper[4934]: E1002 09:49:21.913804 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:49:21 crc kubenswrapper[4934]: E1002 09:49:21.913924 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.977362 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.977395 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.977404 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.977417 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:21 crc kubenswrapper[4934]: I1002 09:49:21.977428 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:21Z","lastTransitionTime":"2025-10-02T09:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.059780 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r"] Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.060239 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.062680 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.063031 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.073105 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.079859 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.079896 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.079907 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.079926 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.079939 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:22Z","lastTransitionTime":"2025-10-02T09:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.084831 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.098079 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.108786 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.118470 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.132434 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.138499 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-th4v6_1d4841b5-0469-461e-875c-25b9fe848141/ovnkube-controller/1.log" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.139081 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-th4v6_1d4841b5-0469-461e-875c-25b9fe848141/ovnkube-controller/0.log" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.142379 4934 generic.go:334] "Generic (PLEG): container finished" podID="1d4841b5-0469-461e-875c-25b9fe848141" containerID="bb5e7ad81f0c3db0a6f84805ce7a8353ac86a75854aef7f2dcc4c0595ef8bf9e" exitCode=1 Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.142421 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" event={"ID":"1d4841b5-0469-461e-875c-25b9fe848141","Type":"ContainerDied","Data":"bb5e7ad81f0c3db0a6f84805ce7a8353ac86a75854aef7f2dcc4c0595ef8bf9e"} Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.142459 4934 scope.go:117] "RemoveContainer" containerID="cc415592d60854b2615100fa4a982146a0b2b770d6fe9e4a61065d13cddb75dc" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.143202 4934 scope.go:117] "RemoveContainer" containerID="bb5e7ad81f0c3db0a6f84805ce7a8353ac86a75854aef7f2dcc4c0595ef8bf9e" Oct 02 09:49:22 crc kubenswrapper[4934]: E1002 09:49:22.143537 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-th4v6_openshift-ovn-kubernetes(1d4841b5-0469-461e-875c-25b9fe848141)\"" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" podUID="1d4841b5-0469-461e-875c-25b9fe848141" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.147055 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.157827 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.167950 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.182360 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.182397 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.182406 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.182419 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.182428 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:22Z","lastTransitionTime":"2025-10-02T09:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.186217 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc415592d60854b2615100fa4a982146a0b2b770d6fe9e4a61065d13cddb75dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc415592d60854b2615100fa4a982146a0b2b770d6fe9e4a61065d13cddb75dc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"message\\\":\\\"kAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1002 09:49:20.272194 6200 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:20.272253 6200 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:49:20.272486 6200 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:49:20.272672 6200 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:20.272711 6200 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:49:20.272883 6200 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:49:20.273167 6200 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.195322 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bpk6\" (UniqueName: \"kubernetes.io/projected/3c9999d1-d367-45d6-8c41-c3bc3ab23df4-kube-api-access-7bpk6\") pod \"ovnkube-control-plane-749d76644c-ffb9r\" (UID: \"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.195385 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3c9999d1-d367-45d6-8c41-c3bc3ab23df4-env-overrides\") pod \"ovnkube-control-plane-749d76644c-ffb9r\" (UID: \"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.195403 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3c9999d1-d367-45d6-8c41-c3bc3ab23df4-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-ffb9r\" (UID: \"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.195430 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3c9999d1-d367-45d6-8c41-c3bc3ab23df4-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-ffb9r\" (UID: \"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.197829 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ffb9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.226725 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.261417 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851e2f96cda63332af97a90a26088ecdd7fcf2d0498b53e5d542aa532b048e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.273568 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.284070 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.284182 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.284238 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.284249 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.284264 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.284277 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:22Z","lastTransitionTime":"2025-10-02T09:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.296358 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3c9999d1-d367-45d6-8c41-c3bc3ab23df4-env-overrides\") pod \"ovnkube-control-plane-749d76644c-ffb9r\" (UID: \"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.296398 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3c9999d1-d367-45d6-8c41-c3bc3ab23df4-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-ffb9r\" (UID: \"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.296455 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3c9999d1-d367-45d6-8c41-c3bc3ab23df4-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-ffb9r\" (UID: \"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.296501 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bpk6\" (UniqueName: \"kubernetes.io/projected/3c9999d1-d367-45d6-8c41-c3bc3ab23df4-kube-api-access-7bpk6\") pod \"ovnkube-control-plane-749d76644c-ffb9r\" (UID: \"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.297009 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3c9999d1-d367-45d6-8c41-c3bc3ab23df4-env-overrides\") pod \"ovnkube-control-plane-749d76644c-ffb9r\" (UID: \"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.297345 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3c9999d1-d367-45d6-8c41-c3bc3ab23df4-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-ffb9r\" (UID: \"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.298173 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.303847 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3c9999d1-d367-45d6-8c41-c3bc3ab23df4-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-ffb9r\" (UID: \"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.313939 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bpk6\" (UniqueName: \"kubernetes.io/projected/3c9999d1-d367-45d6-8c41-c3bc3ab23df4-kube-api-access-7bpk6\") pod \"ovnkube-control-plane-749d76644c-ffb9r\" (UID: \"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.317645 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.332893 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851e2f96cda63332af97a90a26088ecdd7fcf2d0498b53e5d542aa532b048e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.344132 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.361791 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.371671 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.374452 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.385070 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: W1002 09:49:22.385502 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c9999d1_d367_45d6_8c41_c3bc3ab23df4.slice/crio-6ab5848c676bed726aa5cd98e318d91b396349a912960da094849df04bd1e25d WatchSource:0}: Error finding container 6ab5848c676bed726aa5cd98e318d91b396349a912960da094849df04bd1e25d: Status 404 returned error can't find the container with id 6ab5848c676bed726aa5cd98e318d91b396349a912960da094849df04bd1e25d Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.386791 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.386836 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.386851 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.386872 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.386888 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:22Z","lastTransitionTime":"2025-10-02T09:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.398817 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.411692 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.422181 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.435732 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.451182 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.461058 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.473463 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.488944 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.489306 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.489316 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.489331 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.489351 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:22Z","lastTransitionTime":"2025-10-02T09:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.492017 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5e7ad81f0c3db0a6f84805ce7a8353ac86a75854aef7f2dcc4c0595ef8bf9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc415592d60854b2615100fa4a982146a0b2b770d6fe9e4a61065d13cddb75dc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"message\\\":\\\"kAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1002 09:49:20.272194 6200 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:20.272253 6200 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:49:20.272486 6200 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:49:20.272672 6200 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:20.272711 6200 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:49:20.272883 6200 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:49:20.273167 6200 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5e7ad81f0c3db0a6f84805ce7a8353ac86a75854aef7f2dcc4c0595ef8bf9e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:21Z\\\",\\\"message\\\":\\\"penshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:21.916104 6365 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:49:21.915440 6365 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:49:21.916212 6365 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:21.916628 6365 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:21.917211 6365 factory.go:656] Stopping watch factory\\\\nI1002 09:49:21.946642 6365 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1002 09:49:21.946678 6365 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1002 09:49:21.946743 6365 ovnkube.go:599] Stopped ovnkube\\\\nI1002 09:49:21.946777 6365 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 09:49:21.946858 6365 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.503543 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ffb9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.516768 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:22Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.593371 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.593396 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.593405 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.593417 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.593427 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:22Z","lastTransitionTime":"2025-10-02T09:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.696519 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.696568 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.696707 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.696727 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.696739 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:22Z","lastTransitionTime":"2025-10-02T09:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.798725 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.798770 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.798782 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.798799 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.798813 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:22Z","lastTransitionTime":"2025-10-02T09:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.901183 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.901232 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.901242 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.901260 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:22 crc kubenswrapper[4934]: I1002 09:49:22.901271 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:22Z","lastTransitionTime":"2025-10-02T09:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.003611 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.003664 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.003676 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.003695 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.003707 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:23Z","lastTransitionTime":"2025-10-02T09:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.105815 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.105881 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.105893 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.105911 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.105923 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:23Z","lastTransitionTime":"2025-10-02T09:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.147197 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" event={"ID":"3c9999d1-d367-45d6-8c41-c3bc3ab23df4","Type":"ContainerStarted","Data":"87fc3a7a93196f00272230d3c06af1b408d42a806a5e65fbb3cf131ddfb37077"} Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.147253 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" event={"ID":"3c9999d1-d367-45d6-8c41-c3bc3ab23df4","Type":"ContainerStarted","Data":"106b1f23a09ef32bc00222f00777a5a8c211b61c6f3d9fc4b67d7708f8091bc2"} Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.147264 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" event={"ID":"3c9999d1-d367-45d6-8c41-c3bc3ab23df4","Type":"ContainerStarted","Data":"6ab5848c676bed726aa5cd98e318d91b396349a912960da094849df04bd1e25d"} Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.148864 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-th4v6_1d4841b5-0469-461e-875c-25b9fe848141/ovnkube-controller/1.log" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.163640 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851e2f96cda63332af97a90a26088ecdd7fcf2d0498b53e5d542aa532b048e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.188328 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.205997 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.207827 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.207856 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.207864 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.207879 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.207888 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:23Z","lastTransitionTime":"2025-10-02T09:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.226783 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.237283 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.247403 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.257377 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.266218 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.273924 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.283269 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.291518 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.303876 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.310419 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.310461 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.310476 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.310494 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.310507 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:23Z","lastTransitionTime":"2025-10-02T09:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.331254 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5e7ad81f0c3db0a6f84805ce7a8353ac86a75854aef7f2dcc4c0595ef8bf9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc415592d60854b2615100fa4a982146a0b2b770d6fe9e4a61065d13cddb75dc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"message\\\":\\\"kAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1002 09:49:20.272194 6200 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:20.272253 6200 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:49:20.272486 6200 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:49:20.272672 6200 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:20.272711 6200 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:49:20.272883 6200 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:49:20.273167 6200 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5e7ad81f0c3db0a6f84805ce7a8353ac86a75854aef7f2dcc4c0595ef8bf9e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:21Z\\\",\\\"message\\\":\\\"penshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:21.916104 6365 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:49:21.915440 6365 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:49:21.916212 6365 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:21.916628 6365 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:21.917211 6365 factory.go:656] Stopping watch factory\\\\nI1002 09:49:21.946642 6365 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1002 09:49:21.946678 6365 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1002 09:49:21.946743 6365 ovnkube.go:599] Stopped ovnkube\\\\nI1002 09:49:21.946777 6365 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 09:49:21.946858 6365 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.344651 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://106b1f23a09ef32bc00222f00777a5a8c211b61c6f3d9fc4b67d7708f8091bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fc3a7a93196f00272230d3c06af1b408d42a806a5e65fbb3cf131ddfb37077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ffb9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.359246 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.368970 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.412560 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.412609 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.412622 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.412636 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.412646 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:23Z","lastTransitionTime":"2025-10-02T09:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.515521 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.515563 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.515587 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.515602 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.515611 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:23Z","lastTransitionTime":"2025-10-02T09:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.617658 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.617701 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.617712 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.617728 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.617739 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:23Z","lastTransitionTime":"2025-10-02T09:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.720372 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.720422 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.720432 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.720447 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.720456 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:23Z","lastTransitionTime":"2025-10-02T09:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.823207 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.823285 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.823297 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.823316 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.823327 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:23Z","lastTransitionTime":"2025-10-02T09:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.886519 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-n7cgz"] Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.887020 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:23 crc kubenswrapper[4934]: E1002 09:49:23.887093 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.897947 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://106b1f23a09ef32bc00222f00777a5a8c211b61c6f3d9fc4b67d7708f8091bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fc3a7a93196f00272230d3c06af1b408d42a806a5e65fbb3cf131ddfb37077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ffb9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.911124 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:23 crc kubenswrapper[4934]: E1002 09:49:23.911316 4934 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:49:23 crc kubenswrapper[4934]: E1002 09:49:23.911375 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:49:39.911357902 +0000 UTC m=+51.663999424 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.913119 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.913126 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:23 crc kubenswrapper[4934]: E1002 09:49:23.913302 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:49:23 crc kubenswrapper[4934]: E1002 09:49:23.913383 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.913149 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:23 crc kubenswrapper[4934]: E1002 09:49:23.913558 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.918017 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.925242 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.925287 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.925301 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.925323 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.925338 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:23Z","lastTransitionTime":"2025-10-02T09:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.930030 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.938059 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.946383 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.961905 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5e7ad81f0c3db0a6f84805ce7a8353ac86a75854aef7f2dcc4c0595ef8bf9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc415592d60854b2615100fa4a982146a0b2b770d6fe9e4a61065d13cddb75dc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"message\\\":\\\"kAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1002 09:49:20.272194 6200 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:20.272253 6200 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:49:20.272486 6200 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:49:20.272672 6200 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:20.272711 6200 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:49:20.272883 6200 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:49:20.273167 6200 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5e7ad81f0c3db0a6f84805ce7a8353ac86a75854aef7f2dcc4c0595ef8bf9e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:21Z\\\",\\\"message\\\":\\\"penshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:21.916104 6365 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:49:21.915440 6365 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:49:21.916212 6365 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:21.916628 6365 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:21.917211 6365 factory.go:656] Stopping watch factory\\\\nI1002 09:49:21.946642 6365 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1002 09:49:21.946678 6365 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1002 09:49:21.946743 6365 ovnkube.go:599] Stopped ovnkube\\\\nI1002 09:49:21.946777 6365 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 09:49:21.946858 6365 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.982619 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:23 crc kubenswrapper[4934]: I1002 09:49:23.996190 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851e2f96cda63332af97a90a26088ecdd7fcf2d0498b53e5d542aa532b048e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:23Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.008591 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.012098 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.012201 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-metrics-certs\") pod \"network-metrics-daemon-n7cgz\" (UID: \"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\") " pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.012240 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:24 crc kubenswrapper[4934]: E1002 09:49:24.012280 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:49:40.01225506 +0000 UTC m=+51.764896632 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:49:24 crc kubenswrapper[4934]: E1002 09:49:24.012357 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:49:24 crc kubenswrapper[4934]: E1002 09:49:24.012380 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:49:24 crc kubenswrapper[4934]: E1002 09:49:24.012393 4934 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.012354 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:24 crc kubenswrapper[4934]: E1002 09:49:24.012446 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 09:49:40.012428134 +0000 UTC m=+51.765069666 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:49:24 crc kubenswrapper[4934]: E1002 09:49:24.012454 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:49:24 crc kubenswrapper[4934]: E1002 09:49:24.012557 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:49:24 crc kubenswrapper[4934]: E1002 09:49:24.012598 4934 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:49:24 crc kubenswrapper[4934]: E1002 09:49:24.012628 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 09:49:40.012620538 +0000 UTC m=+51.765262060 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.012545 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvdrf\" (UniqueName: \"kubernetes.io/projected/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-kube-api-access-lvdrf\") pod \"network-metrics-daemon-n7cgz\" (UID: \"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\") " pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.012658 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:24 crc kubenswrapper[4934]: E1002 09:49:24.012711 4934 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:49:24 crc kubenswrapper[4934]: E1002 09:49:24.012750 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:49:40.012737491 +0000 UTC m=+51.765379003 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.020987 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.028549 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.028627 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.028640 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.028658 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.028669 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:24Z","lastTransitionTime":"2025-10-02T09:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.034696 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.046763 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7cgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7cgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.057233 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.067939 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.080837 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.094820 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.107965 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.113141 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvdrf\" (UniqueName: \"kubernetes.io/projected/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-kube-api-access-lvdrf\") pod \"network-metrics-daemon-n7cgz\" (UID: \"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\") " pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.113185 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-metrics-certs\") pod \"network-metrics-daemon-n7cgz\" (UID: \"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\") " pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:24 crc kubenswrapper[4934]: E1002 09:49:24.113308 4934 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:49:24 crc kubenswrapper[4934]: E1002 09:49:24.113362 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-metrics-certs podName:0cbef3ee-c66f-47fa-94d3-5ce9892c403f nodeName:}" failed. No retries permitted until 2025-10-02 09:49:24.613344903 +0000 UTC m=+36.365986435 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-metrics-certs") pod "network-metrics-daemon-n7cgz" (UID: "0cbef3ee-c66f-47fa-94d3-5ce9892c403f") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.131046 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.131084 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.131093 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.131109 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.131119 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:24Z","lastTransitionTime":"2025-10-02T09:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.132568 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvdrf\" (UniqueName: \"kubernetes.io/projected/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-kube-api-access-lvdrf\") pod \"network-metrics-daemon-n7cgz\" (UID: \"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\") " pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.235071 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.235136 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.235148 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.235167 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.235181 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:24Z","lastTransitionTime":"2025-10-02T09:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.338292 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.338341 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.338358 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.338379 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.338392 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:24Z","lastTransitionTime":"2025-10-02T09:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.440829 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.440876 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.440888 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.440907 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.440919 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:24Z","lastTransitionTime":"2025-10-02T09:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.543919 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.543978 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.543994 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.544016 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.544034 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:24Z","lastTransitionTime":"2025-10-02T09:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.619877 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-metrics-certs\") pod \"network-metrics-daemon-n7cgz\" (UID: \"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\") " pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:24 crc kubenswrapper[4934]: E1002 09:49:24.620049 4934 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:49:24 crc kubenswrapper[4934]: E1002 09:49:24.620128 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-metrics-certs podName:0cbef3ee-c66f-47fa-94d3-5ce9892c403f nodeName:}" failed. No retries permitted until 2025-10-02 09:49:25.620105398 +0000 UTC m=+37.372746950 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-metrics-certs") pod "network-metrics-daemon-n7cgz" (UID: "0cbef3ee-c66f-47fa-94d3-5ce9892c403f") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.646530 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.646600 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.646626 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.646645 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.646658 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:24Z","lastTransitionTime":"2025-10-02T09:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.748962 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.749046 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.749070 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.749095 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.749114 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:24Z","lastTransitionTime":"2025-10-02T09:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.851365 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.851406 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.851418 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.851435 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.851448 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:24Z","lastTransitionTime":"2025-10-02T09:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.922873 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.922911 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.922923 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.922942 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.922953 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:24Z","lastTransitionTime":"2025-10-02T09:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:24 crc kubenswrapper[4934]: E1002 09:49:24.934660 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.937980 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.938019 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.938028 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.938043 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.938051 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:24Z","lastTransitionTime":"2025-10-02T09:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:24 crc kubenswrapper[4934]: E1002 09:49:24.947859 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.950773 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.950816 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.950828 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.950845 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.950856 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:24Z","lastTransitionTime":"2025-10-02T09:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:24 crc kubenswrapper[4934]: E1002 09:49:24.961925 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.965361 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.965396 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.965407 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.965423 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.965433 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:24Z","lastTransitionTime":"2025-10-02T09:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:24 crc kubenswrapper[4934]: E1002 09:49:24.976510 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.979163 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.979220 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.979253 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.979267 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.979277 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:24Z","lastTransitionTime":"2025-10-02T09:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:24 crc kubenswrapper[4934]: E1002 09:49:24.994537 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:24Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:24 crc kubenswrapper[4934]: E1002 09:49:24.994771 4934 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.996545 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.996603 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.996613 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.996627 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:24 crc kubenswrapper[4934]: I1002 09:49:24.996637 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:24Z","lastTransitionTime":"2025-10-02T09:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.099706 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.099754 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.099765 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.099781 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.099794 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:25Z","lastTransitionTime":"2025-10-02T09:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.202134 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.202167 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.202178 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.202194 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.202205 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:25Z","lastTransitionTime":"2025-10-02T09:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.304162 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.304206 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.304217 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.304234 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.304245 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:25Z","lastTransitionTime":"2025-10-02T09:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.406345 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.406371 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.406379 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.406392 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.406400 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:25Z","lastTransitionTime":"2025-10-02T09:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.508483 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.508523 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.508532 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.508546 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.508555 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:25Z","lastTransitionTime":"2025-10-02T09:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.610948 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.611003 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.611022 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.611044 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.611063 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:25Z","lastTransitionTime":"2025-10-02T09:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.636642 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-metrics-certs\") pod \"network-metrics-daemon-n7cgz\" (UID: \"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\") " pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:25 crc kubenswrapper[4934]: E1002 09:49:25.636770 4934 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:49:25 crc kubenswrapper[4934]: E1002 09:49:25.636825 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-metrics-certs podName:0cbef3ee-c66f-47fa-94d3-5ce9892c403f nodeName:}" failed. No retries permitted until 2025-10-02 09:49:27.636808481 +0000 UTC m=+39.389450003 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-metrics-certs") pod "network-metrics-daemon-n7cgz" (UID: "0cbef3ee-c66f-47fa-94d3-5ce9892c403f") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.713931 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.714000 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.714013 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.714034 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.714046 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:25Z","lastTransitionTime":"2025-10-02T09:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.816966 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.817032 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.817046 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.817080 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.817094 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:25Z","lastTransitionTime":"2025-10-02T09:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.912557 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.912613 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.912627 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.912636 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:25 crc kubenswrapper[4934]: E1002 09:49:25.912789 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:49:25 crc kubenswrapper[4934]: E1002 09:49:25.912853 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:49:25 crc kubenswrapper[4934]: E1002 09:49:25.912938 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:49:25 crc kubenswrapper[4934]: E1002 09:49:25.913116 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.920960 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.921032 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.921043 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.921057 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:25 crc kubenswrapper[4934]: I1002 09:49:25.921066 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:25Z","lastTransitionTime":"2025-10-02T09:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.023995 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.024063 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.024086 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.024119 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.024139 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:26Z","lastTransitionTime":"2025-10-02T09:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.126978 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.127030 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.127046 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.127062 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.127076 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:26Z","lastTransitionTime":"2025-10-02T09:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.229303 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.229348 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.229359 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.229373 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.229387 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:26Z","lastTransitionTime":"2025-10-02T09:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.331406 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.331461 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.331478 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.331499 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.331516 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:26Z","lastTransitionTime":"2025-10-02T09:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.433833 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.433875 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.433887 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.433902 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.433914 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:26Z","lastTransitionTime":"2025-10-02T09:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.535921 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.535951 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.535972 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.535986 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.535995 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:26Z","lastTransitionTime":"2025-10-02T09:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.638625 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.638669 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.638680 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.638699 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.638712 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:26Z","lastTransitionTime":"2025-10-02T09:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.741322 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.741366 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.741390 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.741409 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.741423 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:26Z","lastTransitionTime":"2025-10-02T09:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.844182 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.844244 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.844257 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.844294 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.844306 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:26Z","lastTransitionTime":"2025-10-02T09:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.914171 4934 scope.go:117] "RemoveContainer" containerID="a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.946865 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.947213 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.947224 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.947240 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:26 crc kubenswrapper[4934]: I1002 09:49:26.947249 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:26Z","lastTransitionTime":"2025-10-02T09:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.049599 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.049644 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.049656 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.049670 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.049682 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:27Z","lastTransitionTime":"2025-10-02T09:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.152725 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.152788 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.152800 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.152818 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.152831 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:27Z","lastTransitionTime":"2025-10-02T09:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.166121 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.167803 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"27680347ccabc1c590a7ffc59feeb56ad6123a3ae83806a893ac9283d4c6b5ac"} Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.168187 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.187139 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5e7ad81f0c3db0a6f84805ce7a8353ac86a75854aef7f2dcc4c0595ef8bf9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc415592d60854b2615100fa4a982146a0b2b770d6fe9e4a61065d13cddb75dc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"message\\\":\\\"kAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1002 09:49:20.272194 6200 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:20.272253 6200 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:49:20.272486 6200 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:49:20.272672 6200 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:20.272711 6200 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:49:20.272883 6200 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:49:20.273167 6200 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5e7ad81f0c3db0a6f84805ce7a8353ac86a75854aef7f2dcc4c0595ef8bf9e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:21Z\\\",\\\"message\\\":\\\"penshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:21.916104 6365 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:49:21.915440 6365 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:49:21.916212 6365 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:21.916628 6365 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:21.917211 6365 factory.go:656] Stopping watch factory\\\\nI1002 09:49:21.946642 6365 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1002 09:49:21.946678 6365 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1002 09:49:21.946743 6365 ovnkube.go:599] Stopped ovnkube\\\\nI1002 09:49:21.946777 6365 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 09:49:21.946858 6365 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.197646 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://106b1f23a09ef32bc00222f00777a5a8c211b61c6f3d9fc4b67d7708f8091bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fc3a7a93196f00272230d3c06af1b408d42a806a5e65fbb3cf131ddfb37077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ffb9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.213077 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27680347ccabc1c590a7ffc59feeb56ad6123a3ae83806a893ac9283d4c6b5ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.225597 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.234357 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.244734 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.255792 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.255826 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.255834 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.255847 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.255856 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:27Z","lastTransitionTime":"2025-10-02T09:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.261679 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.274490 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851e2f96cda63332af97a90a26088ecdd7fcf2d0498b53e5d542aa532b048e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.285113 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.295081 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.308672 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.320317 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7cgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7cgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.331705 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.341535 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.352683 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.358123 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.358158 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.358168 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.358184 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.358195 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:27Z","lastTransitionTime":"2025-10-02T09:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.363741 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.373707 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:27Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.461654 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.461723 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.461746 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.461790 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.461813 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:27Z","lastTransitionTime":"2025-10-02T09:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.564924 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.564998 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.565014 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.565040 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.565057 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:27Z","lastTransitionTime":"2025-10-02T09:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.655457 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-metrics-certs\") pod \"network-metrics-daemon-n7cgz\" (UID: \"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\") " pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:27 crc kubenswrapper[4934]: E1002 09:49:27.655633 4934 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:49:27 crc kubenswrapper[4934]: E1002 09:49:27.655693 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-metrics-certs podName:0cbef3ee-c66f-47fa-94d3-5ce9892c403f nodeName:}" failed. No retries permitted until 2025-10-02 09:49:31.655675481 +0000 UTC m=+43.408317003 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-metrics-certs") pod "network-metrics-daemon-n7cgz" (UID: "0cbef3ee-c66f-47fa-94d3-5ce9892c403f") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.667516 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.667552 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.667559 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.667572 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.667617 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:27Z","lastTransitionTime":"2025-10-02T09:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.769849 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.769897 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.769909 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.769924 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.769933 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:27Z","lastTransitionTime":"2025-10-02T09:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.872064 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.872107 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.872121 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.872140 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.872154 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:27Z","lastTransitionTime":"2025-10-02T09:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.912678 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.912715 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.912723 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.912693 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:27 crc kubenswrapper[4934]: E1002 09:49:27.912813 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:49:27 crc kubenswrapper[4934]: E1002 09:49:27.912908 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:49:27 crc kubenswrapper[4934]: E1002 09:49:27.912973 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:49:27 crc kubenswrapper[4934]: E1002 09:49:27.913036 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.974696 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.974746 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.974759 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.974779 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:27 crc kubenswrapper[4934]: I1002 09:49:27.974795 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:27Z","lastTransitionTime":"2025-10-02T09:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.077498 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.077550 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.077569 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.077631 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.077647 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:28Z","lastTransitionTime":"2025-10-02T09:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.180351 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.180398 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.180407 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.180422 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.180432 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:28Z","lastTransitionTime":"2025-10-02T09:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.283646 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.283708 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.283728 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.283777 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.283805 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:28Z","lastTransitionTime":"2025-10-02T09:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.387380 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.387459 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.387483 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.387514 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.387536 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:28Z","lastTransitionTime":"2025-10-02T09:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.490613 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.490655 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.490669 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.490685 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.490697 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:28Z","lastTransitionTime":"2025-10-02T09:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.598186 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.598302 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.598319 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.598343 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.598361 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:28Z","lastTransitionTime":"2025-10-02T09:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.701893 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.701934 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.701961 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.701978 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.701989 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:28Z","lastTransitionTime":"2025-10-02T09:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.805006 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.805033 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.805041 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.805055 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.805064 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:28Z","lastTransitionTime":"2025-10-02T09:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.908327 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.908368 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.908382 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.908400 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.908412 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:28Z","lastTransitionTime":"2025-10-02T09:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.932887 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:28Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.944121 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7cgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7cgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:28Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.958243 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:28Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.970913 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:28Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:28 crc kubenswrapper[4934]: I1002 09:49:28.990403 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:28Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.007881 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.010876 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.010932 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.010955 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.010986 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.011008 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:29Z","lastTransitionTime":"2025-10-02T09:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.029196 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.040773 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.056373 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.068784 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.082441 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.102182 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5e7ad81f0c3db0a6f84805ce7a8353ac86a75854aef7f2dcc4c0595ef8bf9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc415592d60854b2615100fa4a982146a0b2b770d6fe9e4a61065d13cddb75dc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"message\\\":\\\"kAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1002 09:49:20.272194 6200 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:20.272253 6200 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:49:20.272486 6200 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:49:20.272672 6200 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:20.272711 6200 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:49:20.272883 6200 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:49:20.273167 6200 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5e7ad81f0c3db0a6f84805ce7a8353ac86a75854aef7f2dcc4c0595ef8bf9e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:21Z\\\",\\\"message\\\":\\\"penshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:21.916104 6365 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:49:21.915440 6365 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:49:21.916212 6365 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:21.916628 6365 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:21.917211 6365 factory.go:656] Stopping watch factory\\\\nI1002 09:49:21.946642 6365 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1002 09:49:21.946678 6365 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1002 09:49:21.946743 6365 ovnkube.go:599] Stopped ovnkube\\\\nI1002 09:49:21.946777 6365 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 09:49:21.946858 6365 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.113886 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.113931 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.113941 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.113956 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.113967 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:29Z","lastTransitionTime":"2025-10-02T09:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.117557 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://106b1f23a09ef32bc00222f00777a5a8c211b61c6f3d9fc4b67d7708f8091bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fc3a7a93196f00272230d3c06af1b408d42a806a5e65fbb3cf131ddfb37077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ffb9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.135917 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27680347ccabc1c590a7ffc59feeb56ad6123a3ae83806a893ac9283d4c6b5ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.152418 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.170561 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851e2f96cda63332af97a90a26088ecdd7fcf2d0498b53e5d542aa532b048e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.194962 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.216684 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.216729 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.216741 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.216757 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.216769 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:29Z","lastTransitionTime":"2025-10-02T09:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.319907 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.319959 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.319976 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.319999 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.320017 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:29Z","lastTransitionTime":"2025-10-02T09:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.423078 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.423119 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.423128 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.423150 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.423160 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:29Z","lastTransitionTime":"2025-10-02T09:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.526385 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.526417 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.526426 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.526441 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.526449 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:29Z","lastTransitionTime":"2025-10-02T09:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.628610 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.628643 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.628651 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.628682 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.628691 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:29Z","lastTransitionTime":"2025-10-02T09:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.731066 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.731124 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.731140 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.731163 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.731179 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:29Z","lastTransitionTime":"2025-10-02T09:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.835018 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.835098 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.835120 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.835153 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.835188 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:29Z","lastTransitionTime":"2025-10-02T09:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.912656 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.912769 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:29 crc kubenswrapper[4934]: E1002 09:49:29.912819 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.912835 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.912877 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:29 crc kubenswrapper[4934]: E1002 09:49:29.913006 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:49:29 crc kubenswrapper[4934]: E1002 09:49:29.913175 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:49:29 crc kubenswrapper[4934]: E1002 09:49:29.913325 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.937777 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.937827 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.937838 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.937856 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:29 crc kubenswrapper[4934]: I1002 09:49:29.937868 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:29Z","lastTransitionTime":"2025-10-02T09:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.040885 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.040936 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.040977 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.041000 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.041015 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:30Z","lastTransitionTime":"2025-10-02T09:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.144544 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.144618 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.144634 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.144653 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.144667 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:30Z","lastTransitionTime":"2025-10-02T09:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.247059 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.247127 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.247145 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.247179 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.247197 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:30Z","lastTransitionTime":"2025-10-02T09:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.349701 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.350002 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.350142 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.350267 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.350339 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:30Z","lastTransitionTime":"2025-10-02T09:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.452980 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.453015 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.453023 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.453054 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.453064 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:30Z","lastTransitionTime":"2025-10-02T09:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.556247 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.556308 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.556321 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.556336 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.556347 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:30Z","lastTransitionTime":"2025-10-02T09:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.658665 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.658744 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.658776 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.658804 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.658826 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:30Z","lastTransitionTime":"2025-10-02T09:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.761756 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.762038 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.762090 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.762124 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.762149 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:30Z","lastTransitionTime":"2025-10-02T09:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.864411 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.864482 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.864500 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.864523 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.864540 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:30Z","lastTransitionTime":"2025-10-02T09:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.966947 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.967008 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.967024 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.967048 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:30 crc kubenswrapper[4934]: I1002 09:49:30.967063 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:30Z","lastTransitionTime":"2025-10-02T09:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.069642 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.069681 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.069689 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.069703 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.069714 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:31Z","lastTransitionTime":"2025-10-02T09:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.172222 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.172260 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.172272 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.172289 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.172300 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:31Z","lastTransitionTime":"2025-10-02T09:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.274734 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.274780 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.274790 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.274806 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.274816 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:31Z","lastTransitionTime":"2025-10-02T09:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.377914 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.377966 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.377983 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.378008 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.378025 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:31Z","lastTransitionTime":"2025-10-02T09:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.481451 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.481499 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.481522 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.481540 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.481551 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:31Z","lastTransitionTime":"2025-10-02T09:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.584691 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.584730 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.584738 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.584756 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.584767 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:31Z","lastTransitionTime":"2025-10-02T09:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.687617 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.687671 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.687685 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.687703 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.687714 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:31Z","lastTransitionTime":"2025-10-02T09:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.698224 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-metrics-certs\") pod \"network-metrics-daemon-n7cgz\" (UID: \"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\") " pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:31 crc kubenswrapper[4934]: E1002 09:49:31.698348 4934 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:49:31 crc kubenswrapper[4934]: E1002 09:49:31.698402 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-metrics-certs podName:0cbef3ee-c66f-47fa-94d3-5ce9892c403f nodeName:}" failed. No retries permitted until 2025-10-02 09:49:39.698386828 +0000 UTC m=+51.451028350 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-metrics-certs") pod "network-metrics-daemon-n7cgz" (UID: "0cbef3ee-c66f-47fa-94d3-5ce9892c403f") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.790307 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.790344 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.790355 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.790371 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.790381 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:31Z","lastTransitionTime":"2025-10-02T09:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.892385 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.892420 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.892428 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.892441 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.892450 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:31Z","lastTransitionTime":"2025-10-02T09:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.912462 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.912564 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.912564 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:31 crc kubenswrapper[4934]: E1002 09:49:31.912667 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:49:31 crc kubenswrapper[4934]: E1002 09:49:31.912808 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.912829 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:31 crc kubenswrapper[4934]: E1002 09:49:31.912961 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:49:31 crc kubenswrapper[4934]: E1002 09:49:31.913050 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.994860 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.994935 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.994955 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.994979 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:31 crc kubenswrapper[4934]: I1002 09:49:31.994998 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:31Z","lastTransitionTime":"2025-10-02T09:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.098457 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.098513 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.098522 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.098536 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.098546 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:32Z","lastTransitionTime":"2025-10-02T09:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.201238 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.201302 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.201350 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.201366 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.201389 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:32Z","lastTransitionTime":"2025-10-02T09:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.303947 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.303986 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.303999 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.304015 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.304027 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:32Z","lastTransitionTime":"2025-10-02T09:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.406718 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.406753 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.406764 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.406780 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.406791 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:32Z","lastTransitionTime":"2025-10-02T09:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.509050 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.509134 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.509159 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.509192 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.509219 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:32Z","lastTransitionTime":"2025-10-02T09:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.612026 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.612073 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.612090 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.612113 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.612130 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:32Z","lastTransitionTime":"2025-10-02T09:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.715939 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.715999 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.716016 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.716041 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.716058 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:32Z","lastTransitionTime":"2025-10-02T09:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.819813 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.820375 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.820400 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.820421 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.820433 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:32Z","lastTransitionTime":"2025-10-02T09:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.914404 4934 scope.go:117] "RemoveContainer" containerID="bb5e7ad81f0c3db0a6f84805ce7a8353ac86a75854aef7f2dcc4c0595ef8bf9e" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.922550 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.922612 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.922632 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.922665 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.922679 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:32Z","lastTransitionTime":"2025-10-02T09:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.938328 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.956690 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.975387 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:32 crc kubenswrapper[4934]: I1002 09:49:32.990836 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7cgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7cgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:32Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.010055 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.024348 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.025640 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.025678 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.025690 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.025710 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.025724 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:33Z","lastTransitionTime":"2025-10-02T09:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.041043 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.055142 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.066262 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.081322 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27680347ccabc1c590a7ffc59feeb56ad6123a3ae83806a893ac9283d4c6b5ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.096108 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.112226 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.123005 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.128051 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.128088 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.128096 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.128114 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.128124 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:33Z","lastTransitionTime":"2025-10-02T09:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.140334 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb5e7ad81f0c3db0a6f84805ce7a8353ac86a75854aef7f2dcc4c0595ef8bf9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5e7ad81f0c3db0a6f84805ce7a8353ac86a75854aef7f2dcc4c0595ef8bf9e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:21Z\\\",\\\"message\\\":\\\"penshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:21.916104 6365 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:49:21.915440 6365 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:49:21.916212 6365 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:21.916628 6365 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:21.917211 6365 factory.go:656] Stopping watch factory\\\\nI1002 09:49:21.946642 6365 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1002 09:49:21.946678 6365 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1002 09:49:21.946743 6365 ovnkube.go:599] Stopped ovnkube\\\\nI1002 09:49:21.946777 6365 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 09:49:21.946858 6365 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-th4v6_openshift-ovn-kubernetes(1d4841b5-0469-461e-875c-25b9fe848141)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.151689 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://106b1f23a09ef32bc00222f00777a5a8c211b61c6f3d9fc4b67d7708f8091bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fc3a7a93196f00272230d3c06af1b408d42a806a5e65fbb3cf131ddfb37077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ffb9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.169367 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.181659 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851e2f96cda63332af97a90a26088ecdd7fcf2d0498b53e5d542aa532b048e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.190739 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-th4v6_1d4841b5-0469-461e-875c-25b9fe848141/ovnkube-controller/1.log" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.193282 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" event={"ID":"1d4841b5-0469-461e-875c-25b9fe848141","Type":"ContainerStarted","Data":"9bfd39d2fcb3f63e1b033c6db4503d8ce5466e7af8555d0890174db12cca9773"} Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.193406 4934 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.205624 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.215931 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7cgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7cgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.227890 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.231374 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.231400 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.231410 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.231447 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.231456 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:33Z","lastTransitionTime":"2025-10-02T09:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.239874 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.250925 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.262982 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.274032 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.286090 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.298958 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.309831 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.322428 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.333523 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.333568 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.333591 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.333609 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.333619 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:33Z","lastTransitionTime":"2025-10-02T09:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.354274 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bfd39d2fcb3f63e1b033c6db4503d8ce5466e7af8555d0890174db12cca9773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5e7ad81f0c3db0a6f84805ce7a8353ac86a75854aef7f2dcc4c0595ef8bf9e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:21Z\\\",\\\"message\\\":\\\"penshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:21.916104 6365 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:49:21.915440 6365 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:49:21.916212 6365 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:21.916628 6365 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:21.917211 6365 factory.go:656] Stopping watch factory\\\\nI1002 09:49:21.946642 6365 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1002 09:49:21.946678 6365 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1002 09:49:21.946743 6365 ovnkube.go:599] Stopped ovnkube\\\\nI1002 09:49:21.946777 6365 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 09:49:21.946858 6365 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.371222 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://106b1f23a09ef32bc00222f00777a5a8c211b61c6f3d9fc4b67d7708f8091bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fc3a7a93196f00272230d3c06af1b408d42a806a5e65fbb3cf131ddfb37077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ffb9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.384293 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27680347ccabc1c590a7ffc59feeb56ad6123a3ae83806a893ac9283d4c6b5ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.395759 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.408227 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851e2f96cda63332af97a90a26088ecdd7fcf2d0498b53e5d542aa532b048e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.425089 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.435726 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.435760 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.435768 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.435784 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.435794 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:33Z","lastTransitionTime":"2025-10-02T09:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.538197 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.538230 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.538238 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.538251 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.538260 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:33Z","lastTransitionTime":"2025-10-02T09:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.640661 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.640694 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.640702 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.640716 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.640724 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:33Z","lastTransitionTime":"2025-10-02T09:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.743853 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.743903 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.743914 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.743929 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.743939 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:33Z","lastTransitionTime":"2025-10-02T09:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.847173 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.847229 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.847245 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.847270 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.847288 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:33Z","lastTransitionTime":"2025-10-02T09:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.912511 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.912530 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:33 crc kubenswrapper[4934]: E1002 09:49:33.912774 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.912534 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:33 crc kubenswrapper[4934]: E1002 09:49:33.912872 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.912912 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:33 crc kubenswrapper[4934]: E1002 09:49:33.913041 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:49:33 crc kubenswrapper[4934]: E1002 09:49:33.913135 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.949254 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.949320 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.949337 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.949361 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:33 crc kubenswrapper[4934]: I1002 09:49:33.949378 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:33Z","lastTransitionTime":"2025-10-02T09:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.053402 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.053481 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.053501 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.053526 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.053543 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:34Z","lastTransitionTime":"2025-10-02T09:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.156230 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.156302 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.156315 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.156332 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.156345 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:34Z","lastTransitionTime":"2025-10-02T09:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.199158 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-th4v6_1d4841b5-0469-461e-875c-25b9fe848141/ovnkube-controller/2.log" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.200102 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-th4v6_1d4841b5-0469-461e-875c-25b9fe848141/ovnkube-controller/1.log" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.203207 4934 generic.go:334] "Generic (PLEG): container finished" podID="1d4841b5-0469-461e-875c-25b9fe848141" containerID="9bfd39d2fcb3f63e1b033c6db4503d8ce5466e7af8555d0890174db12cca9773" exitCode=1 Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.203256 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" event={"ID":"1d4841b5-0469-461e-875c-25b9fe848141","Type":"ContainerDied","Data":"9bfd39d2fcb3f63e1b033c6db4503d8ce5466e7af8555d0890174db12cca9773"} Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.203304 4934 scope.go:117] "RemoveContainer" containerID="bb5e7ad81f0c3db0a6f84805ce7a8353ac86a75854aef7f2dcc4c0595ef8bf9e" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.204476 4934 scope.go:117] "RemoveContainer" containerID="9bfd39d2fcb3f63e1b033c6db4503d8ce5466e7af8555d0890174db12cca9773" Oct 02 09:49:34 crc kubenswrapper[4934]: E1002 09:49:34.204778 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-th4v6_openshift-ovn-kubernetes(1d4841b5-0469-461e-875c-25b9fe848141)\"" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" podUID="1d4841b5-0469-461e-875c-25b9fe848141" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.216288 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:34Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.235326 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bfd39d2fcb3f63e1b033c6db4503d8ce5466e7af8555d0890174db12cca9773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5e7ad81f0c3db0a6f84805ce7a8353ac86a75854aef7f2dcc4c0595ef8bf9e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:21Z\\\",\\\"message\\\":\\\"penshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:21.916104 6365 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:49:21.915440 6365 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:49:21.916212 6365 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:21.916628 6365 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:21.917211 6365 factory.go:656] Stopping watch factory\\\\nI1002 09:49:21.946642 6365 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1002 09:49:21.946678 6365 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1002 09:49:21.946743 6365 ovnkube.go:599] Stopped ovnkube\\\\nI1002 09:49:21.946777 6365 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 09:49:21.946858 6365 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfd39d2fcb3f63e1b033c6db4503d8ce5466e7af8555d0890174db12cca9773\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:33Z\\\",\\\"message\\\":\\\"tions: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z]\\\\nI1002 09:49:33.730641 6587 services_controller.go:434] Service openshift-operator-lifecycle-manager/packageserver-service retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{packageserver-service openshift-operator-lifecycle-manager a60a1f74-c6ff-4c81-96ae-27ba9796ba61 5485 0 2025-02-23 05:23:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true] map[] [{operators.coreos.com/v1alpha1 ClusterServiceVersion packageserver bbc08db6-5ba4-4fc4-b49d-26331e1e728b 0xc0076f7bcd 0xc0076f7bce}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:5443,Protocol:TCP,Port:5443,TargetPort:{0 5443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: packageserver,},ClusterIP:10.217.4.153,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.153],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:34Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.248132 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://106b1f23a09ef32bc00222f00777a5a8c211b61c6f3d9fc4b67d7708f8091bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fc3a7a93196f00272230d3c06af1b408d42a806a5e65fbb3cf131ddfb37077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ffb9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:34Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.258552 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.258615 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.258625 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.258639 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.258649 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:34Z","lastTransitionTime":"2025-10-02T09:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.263644 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27680347ccabc1c590a7ffc59feeb56ad6123a3ae83806a893ac9283d4c6b5ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:34Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.275142 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:34Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.285428 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:34Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.302493 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:34Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.316247 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851e2f96cda63332af97a90a26088ecdd7fcf2d0498b53e5d542aa532b048e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:34Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.324637 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7cgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7cgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:34Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.338109 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:34Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.349409 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:34Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.359983 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:34Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.360616 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.360732 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.360831 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.360919 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.361012 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:34Z","lastTransitionTime":"2025-10-02T09:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.372804 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:34Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.383639 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:34Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.392061 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:34Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.402812 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:34Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.415439 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:34Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.463563 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.463624 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.463637 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.463653 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.463664 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:34Z","lastTransitionTime":"2025-10-02T09:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.567303 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.568116 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.568289 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.568371 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.568478 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:34Z","lastTransitionTime":"2025-10-02T09:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.670797 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.670844 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.670856 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.670875 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.670889 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:34Z","lastTransitionTime":"2025-10-02T09:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.773514 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.773553 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.773562 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.773595 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.773613 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:34Z","lastTransitionTime":"2025-10-02T09:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.876531 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.876597 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.876609 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.876630 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.876648 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:34Z","lastTransitionTime":"2025-10-02T09:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.979369 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.979408 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.979417 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.979433 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:34 crc kubenswrapper[4934]: I1002 09:49:34.979444 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:34Z","lastTransitionTime":"2025-10-02T09:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.034313 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.034375 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.034395 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.034425 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.034448 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:35Z","lastTransitionTime":"2025-10-02T09:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:35 crc kubenswrapper[4934]: E1002 09:49:35.054737 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:35Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.064787 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.064861 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.064883 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.064913 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.064935 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:35Z","lastTransitionTime":"2025-10-02T09:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:35 crc kubenswrapper[4934]: E1002 09:49:35.079928 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:35Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.084943 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.085182 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.085204 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.085232 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.085254 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:35Z","lastTransitionTime":"2025-10-02T09:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:35 crc kubenswrapper[4934]: E1002 09:49:35.104968 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:35Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.110319 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.110376 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.110387 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.110402 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.110413 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:35Z","lastTransitionTime":"2025-10-02T09:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:35 crc kubenswrapper[4934]: E1002 09:49:35.128366 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:35Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.131964 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.132041 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.132066 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.132131 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.132158 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:35Z","lastTransitionTime":"2025-10-02T09:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:35 crc kubenswrapper[4934]: E1002 09:49:35.148061 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:35Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:35 crc kubenswrapper[4934]: E1002 09:49:35.148249 4934 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.149828 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.149897 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.149920 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.149951 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.149976 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:35Z","lastTransitionTime":"2025-10-02T09:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.208520 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-th4v6_1d4841b5-0469-461e-875c-25b9fe848141/ovnkube-controller/2.log" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.252657 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.252714 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.252727 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.252746 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.252792 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:35Z","lastTransitionTime":"2025-10-02T09:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.355676 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.355721 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.355739 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.355765 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.355782 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:35Z","lastTransitionTime":"2025-10-02T09:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.457682 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.457747 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.457766 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.457789 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.457806 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:35Z","lastTransitionTime":"2025-10-02T09:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.560784 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.561098 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.561205 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.561321 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.561399 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:35Z","lastTransitionTime":"2025-10-02T09:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.664522 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.665330 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.665432 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.665538 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.665671 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:35Z","lastTransitionTime":"2025-10-02T09:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.768978 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.769058 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.769079 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.769106 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.769125 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:35Z","lastTransitionTime":"2025-10-02T09:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.872986 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.873018 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.873028 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.873041 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.873050 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:35Z","lastTransitionTime":"2025-10-02T09:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.912837 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.912955 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:35 crc kubenswrapper[4934]: E1002 09:49:35.913285 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.913336 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.913398 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:35 crc kubenswrapper[4934]: E1002 09:49:35.913474 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:49:35 crc kubenswrapper[4934]: E1002 09:49:35.913415 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:49:35 crc kubenswrapper[4934]: E1002 09:49:35.913693 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.975760 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.975796 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.975805 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.975820 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:35 crc kubenswrapper[4934]: I1002 09:49:35.975831 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:35Z","lastTransitionTime":"2025-10-02T09:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.078223 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.078291 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.078308 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.078332 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.078349 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:36Z","lastTransitionTime":"2025-10-02T09:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.180402 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.180441 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.180450 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.180464 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.180472 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:36Z","lastTransitionTime":"2025-10-02T09:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.283471 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.283522 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.283535 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.283552 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.283563 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:36Z","lastTransitionTime":"2025-10-02T09:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.385367 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.385422 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.385433 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.385452 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.385464 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:36Z","lastTransitionTime":"2025-10-02T09:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.487652 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.487691 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.487701 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.487716 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.487725 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:36Z","lastTransitionTime":"2025-10-02T09:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.590896 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.590959 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.590971 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.591002 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.591018 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:36Z","lastTransitionTime":"2025-10-02T09:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.696423 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.696464 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.696684 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.696702 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.696713 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:36Z","lastTransitionTime":"2025-10-02T09:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.799224 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.799321 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.799337 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.799353 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.799364 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:36Z","lastTransitionTime":"2025-10-02T09:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.902063 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.902093 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.902103 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.902144 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:36 crc kubenswrapper[4934]: I1002 09:49:36.902157 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:36Z","lastTransitionTime":"2025-10-02T09:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.004406 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.004472 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.004487 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.004506 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.004545 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:37Z","lastTransitionTime":"2025-10-02T09:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.107640 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.107720 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.107733 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.107778 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.107793 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:37Z","lastTransitionTime":"2025-10-02T09:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.210837 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.210899 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.210916 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.210938 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.210954 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:37Z","lastTransitionTime":"2025-10-02T09:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.313691 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.313740 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.313752 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.313768 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.313779 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:37Z","lastTransitionTime":"2025-10-02T09:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.415526 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.415608 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.415628 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.415651 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.415666 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:37Z","lastTransitionTime":"2025-10-02T09:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.518523 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.518657 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.518693 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.518726 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.518747 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:37Z","lastTransitionTime":"2025-10-02T09:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.622172 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.622246 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.622282 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.622312 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.622336 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:37Z","lastTransitionTime":"2025-10-02T09:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.725538 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.725672 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.725701 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.725731 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.725749 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:37Z","lastTransitionTime":"2025-10-02T09:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.828060 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.828124 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.828137 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.828154 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.828166 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:37Z","lastTransitionTime":"2025-10-02T09:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.912572 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.912710 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:37 crc kubenswrapper[4934]: E1002 09:49:37.912766 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.912788 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.912710 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:37 crc kubenswrapper[4934]: E1002 09:49:37.912912 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:49:37 crc kubenswrapper[4934]: E1002 09:49:37.913037 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:49:37 crc kubenswrapper[4934]: E1002 09:49:37.913161 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.930317 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.930351 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.930361 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.930377 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:37 crc kubenswrapper[4934]: I1002 09:49:37.930390 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:37Z","lastTransitionTime":"2025-10-02T09:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.033722 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.033764 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.033778 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.033793 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.033802 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:38Z","lastTransitionTime":"2025-10-02T09:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.137186 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.137232 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.137244 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.137263 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.137275 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:38Z","lastTransitionTime":"2025-10-02T09:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.239142 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.239178 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.239186 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.239201 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.239210 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:38Z","lastTransitionTime":"2025-10-02T09:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.341924 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.341969 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.341987 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.342072 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.342093 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:38Z","lastTransitionTime":"2025-10-02T09:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.445194 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.445262 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.445274 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.445294 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.445308 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:38Z","lastTransitionTime":"2025-10-02T09:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.548751 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.548797 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.548809 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.548827 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.548838 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:38Z","lastTransitionTime":"2025-10-02T09:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.652067 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.652115 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.652125 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.652145 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.652157 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:38Z","lastTransitionTime":"2025-10-02T09:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.755294 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.756179 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.756265 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.756354 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.756430 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:38Z","lastTransitionTime":"2025-10-02T09:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.859750 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.860119 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.860336 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.860520 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.860731 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:38Z","lastTransitionTime":"2025-10-02T09:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.931956 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:38Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.946974 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:38Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.958943 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:38Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.963092 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.963129 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.963139 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.963158 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.963171 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:38Z","lastTransitionTime":"2025-10-02T09:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.970561 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:38Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:38 crc kubenswrapper[4934]: I1002 09:49:38.985133 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:38Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.000274 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:38Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.015899 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.036527 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bfd39d2fcb3f63e1b033c6db4503d8ce5466e7af8555d0890174db12cca9773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb5e7ad81f0c3db0a6f84805ce7a8353ac86a75854aef7f2dcc4c0595ef8bf9e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:21Z\\\",\\\"message\\\":\\\"penshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:21.916104 6365 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:49:21.915440 6365 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:49:21.916212 6365 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:21.916628 6365 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:49:21.917211 6365 factory.go:656] Stopping watch factory\\\\nI1002 09:49:21.946642 6365 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1002 09:49:21.946678 6365 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1002 09:49:21.946743 6365 ovnkube.go:599] Stopped ovnkube\\\\nI1002 09:49:21.946777 6365 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 09:49:21.946858 6365 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfd39d2fcb3f63e1b033c6db4503d8ce5466e7af8555d0890174db12cca9773\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:33Z\\\",\\\"message\\\":\\\"tions: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z]\\\\nI1002 09:49:33.730641 6587 services_controller.go:434] Service openshift-operator-lifecycle-manager/packageserver-service retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{packageserver-service openshift-operator-lifecycle-manager a60a1f74-c6ff-4c81-96ae-27ba9796ba61 5485 0 2025-02-23 05:23:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true] map[] [{operators.coreos.com/v1alpha1 ClusterServiceVersion packageserver bbc08db6-5ba4-4fc4-b49d-26331e1e728b 0xc0076f7bcd 0xc0076f7bce}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:5443,Protocol:TCP,Port:5443,TargetPort:{0 5443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: packageserver,},ClusterIP:10.217.4.153,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.153],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.049123 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://106b1f23a09ef32bc00222f00777a5a8c211b61c6f3d9fc4b67d7708f8091bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fc3a7a93196f00272230d3c06af1b408d42a806a5e65fbb3cf131ddfb37077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ffb9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.063844 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27680347ccabc1c590a7ffc59feeb56ad6123a3ae83806a893ac9283d4c6b5ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.065196 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.065228 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.065237 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.065250 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.065260 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:39Z","lastTransitionTime":"2025-10-02T09:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.076353 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.090159 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851e2f96cda63332af97a90a26088ecdd7fcf2d0498b53e5d542aa532b048e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.109025 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.123932 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.134126 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7cgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7cgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.150185 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.166445 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:39Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.168194 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.168256 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.168274 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.168296 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.168314 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:39Z","lastTransitionTime":"2025-10-02T09:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.271499 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.271563 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.271605 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.271624 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.271634 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:39Z","lastTransitionTime":"2025-10-02T09:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.374138 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.374193 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.374204 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.374219 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.374228 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:39Z","lastTransitionTime":"2025-10-02T09:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.477737 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.477812 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.477824 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.477843 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.477860 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:39Z","lastTransitionTime":"2025-10-02T09:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.580287 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.580335 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.580348 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.580369 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.580383 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:39Z","lastTransitionTime":"2025-10-02T09:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.683410 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.683460 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.683471 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.683487 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.683501 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:39Z","lastTransitionTime":"2025-10-02T09:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.781010 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-metrics-certs\") pod \"network-metrics-daemon-n7cgz\" (UID: \"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\") " pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:39 crc kubenswrapper[4934]: E1002 09:49:39.781342 4934 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:49:39 crc kubenswrapper[4934]: E1002 09:49:39.781425 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-metrics-certs podName:0cbef3ee-c66f-47fa-94d3-5ce9892c403f nodeName:}" failed. No retries permitted until 2025-10-02 09:49:55.781400294 +0000 UTC m=+67.534041816 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-metrics-certs") pod "network-metrics-daemon-n7cgz" (UID: "0cbef3ee-c66f-47fa-94d3-5ce9892c403f") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.787816 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.787892 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.787907 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.787931 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.787951 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:39Z","lastTransitionTime":"2025-10-02T09:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.891365 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.891399 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.891408 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.891425 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.891434 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:39Z","lastTransitionTime":"2025-10-02T09:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.912890 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.912981 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.913099 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:39 crc kubenswrapper[4934]: E1002 09:49:39.913087 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:49:39 crc kubenswrapper[4934]: E1002 09:49:39.913348 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.912921 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:39 crc kubenswrapper[4934]: E1002 09:49:39.913472 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:49:39 crc kubenswrapper[4934]: E1002 09:49:39.913604 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.983079 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:39 crc kubenswrapper[4934]: E1002 09:49:39.983299 4934 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:49:39 crc kubenswrapper[4934]: E1002 09:49:39.983400 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:50:11.983377765 +0000 UTC m=+83.736019337 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.993866 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.993928 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.993942 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.993964 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:39 crc kubenswrapper[4934]: I1002 09:49:39.993978 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:39Z","lastTransitionTime":"2025-10-02T09:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.084555 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:49:40 crc kubenswrapper[4934]: E1002 09:49:40.084845 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:12.084803915 +0000 UTC m=+83.837445437 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.084985 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.085034 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.085155 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:40 crc kubenswrapper[4934]: E1002 09:49:40.085265 4934 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:49:40 crc kubenswrapper[4934]: E1002 09:49:40.085304 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:49:40 crc kubenswrapper[4934]: E1002 09:49:40.085344 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:49:40 crc kubenswrapper[4934]: E1002 09:49:40.085349 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:49:40 crc kubenswrapper[4934]: E1002 09:49:40.085370 4934 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:49:40 crc kubenswrapper[4934]: E1002 09:49:40.085374 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:49:40 crc kubenswrapper[4934]: E1002 09:49:40.085537 4934 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:49:40 crc kubenswrapper[4934]: E1002 09:49:40.085357 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:50:12.085331207 +0000 UTC m=+83.837972739 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:49:40 crc kubenswrapper[4934]: E1002 09:49:40.085630 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 09:50:12.085616564 +0000 UTC m=+83.838258086 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:49:40 crc kubenswrapper[4934]: E1002 09:49:40.085674 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 09:50:12.085638654 +0000 UTC m=+83.838280346 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.096986 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.097042 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.097051 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.097070 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.097097 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:40Z","lastTransitionTime":"2025-10-02T09:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.199792 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.199832 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.199846 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.199863 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.199874 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:40Z","lastTransitionTime":"2025-10-02T09:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.302821 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.302879 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.302892 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.302910 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.302922 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:40Z","lastTransitionTime":"2025-10-02T09:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.405819 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.405878 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.405891 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.405915 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.405931 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:40Z","lastTransitionTime":"2025-10-02T09:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.508272 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.508336 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.508354 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.508376 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.508391 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:40Z","lastTransitionTime":"2025-10-02T09:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.611220 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.611291 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.611316 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.611351 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.611374 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:40Z","lastTransitionTime":"2025-10-02T09:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.650730 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.651773 4934 scope.go:117] "RemoveContainer" containerID="9bfd39d2fcb3f63e1b033c6db4503d8ce5466e7af8555d0890174db12cca9773" Oct 02 09:49:40 crc kubenswrapper[4934]: E1002 09:49:40.652017 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-th4v6_openshift-ovn-kubernetes(1d4841b5-0469-461e-875c-25b9fe848141)\"" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" podUID="1d4841b5-0469-461e-875c-25b9fe848141" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.671489 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.689361 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851e2f96cda63332af97a90a26088ecdd7fcf2d0498b53e5d542aa532b048e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.701393 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7cgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7cgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.714316 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.714407 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.714422 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.714444 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.714458 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:40Z","lastTransitionTime":"2025-10-02T09:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.716179 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.730612 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.746726 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.761506 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.775971 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.788239 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.806934 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.817449 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.817505 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.817523 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.817546 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.817562 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:40Z","lastTransitionTime":"2025-10-02T09:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.821056 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.834628 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.859068 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bfd39d2fcb3f63e1b033c6db4503d8ce5466e7af8555d0890174db12cca9773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfd39d2fcb3f63e1b033c6db4503d8ce5466e7af8555d0890174db12cca9773\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:33Z\\\",\\\"message\\\":\\\"tions: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z]\\\\nI1002 09:49:33.730641 6587 services_controller.go:434] Service openshift-operator-lifecycle-manager/packageserver-service retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{packageserver-service openshift-operator-lifecycle-manager a60a1f74-c6ff-4c81-96ae-27ba9796ba61 5485 0 2025-02-23 05:23:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true] map[] [{operators.coreos.com/v1alpha1 ClusterServiceVersion packageserver bbc08db6-5ba4-4fc4-b49d-26331e1e728b 0xc0076f7bcd 0xc0076f7bce}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:5443,Protocol:TCP,Port:5443,TargetPort:{0 5443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: packageserver,},ClusterIP:10.217.4.153,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.153],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-th4v6_openshift-ovn-kubernetes(1d4841b5-0469-461e-875c-25b9fe848141)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.875651 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://106b1f23a09ef32bc00222f00777a5a8c211b61c6f3d9fc4b67d7708f8091bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fc3a7a93196f00272230d3c06af1b408d42a806a5e65fbb3cf131ddfb37077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ffb9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.894949 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27680347ccabc1c590a7ffc59feeb56ad6123a3ae83806a893ac9283d4c6b5ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.913040 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.919951 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.919992 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.920003 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.920019 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.920030 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:40Z","lastTransitionTime":"2025-10-02T09:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:40 crc kubenswrapper[4934]: I1002 09:49:40.923102 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:40Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.022757 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.022799 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.022811 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.022850 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.022871 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:41Z","lastTransitionTime":"2025-10-02T09:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.095187 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.123768 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.125340 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.125363 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.125371 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.125386 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.125398 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:41Z","lastTransitionTime":"2025-10-02T09:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.140089 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851e2f96cda63332af97a90a26088ecdd7fcf2d0498b53e5d542aa532b048e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.155996 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.169869 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.184955 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.199376 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7cgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7cgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.213631 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.226771 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.228438 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.228492 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.228503 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.228521 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.228532 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:41Z","lastTransitionTime":"2025-10-02T09:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.243675 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.256339 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.269685 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.283630 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27680347ccabc1c590a7ffc59feeb56ad6123a3ae83806a893ac9283d4c6b5ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.297603 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.308271 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.322572 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.331414 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.331456 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.331467 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.331485 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.331497 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:41Z","lastTransitionTime":"2025-10-02T09:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.343464 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bfd39d2fcb3f63e1b033c6db4503d8ce5466e7af8555d0890174db12cca9773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfd39d2fcb3f63e1b033c6db4503d8ce5466e7af8555d0890174db12cca9773\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:33Z\\\",\\\"message\\\":\\\"tions: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z]\\\\nI1002 09:49:33.730641 6587 services_controller.go:434] Service openshift-operator-lifecycle-manager/packageserver-service retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{packageserver-service openshift-operator-lifecycle-manager a60a1f74-c6ff-4c81-96ae-27ba9796ba61 5485 0 2025-02-23 05:23:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true] map[] [{operators.coreos.com/v1alpha1 ClusterServiceVersion packageserver bbc08db6-5ba4-4fc4-b49d-26331e1e728b 0xc0076f7bcd 0xc0076f7bce}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:5443,Protocol:TCP,Port:5443,TargetPort:{0 5443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: packageserver,},ClusterIP:10.217.4.153,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.153],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-th4v6_openshift-ovn-kubernetes(1d4841b5-0469-461e-875c-25b9fe848141)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.356110 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://106b1f23a09ef32bc00222f00777a5a8c211b61c6f3d9fc4b67d7708f8091bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fc3a7a93196f00272230d3c06af1b408d42a806a5e65fbb3cf131ddfb37077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ffb9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:41Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.435975 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.436067 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.436090 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.436117 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.436138 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:41Z","lastTransitionTime":"2025-10-02T09:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.538975 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.539032 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.539053 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.539072 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.539086 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:41Z","lastTransitionTime":"2025-10-02T09:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.642434 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.642512 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.642533 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.642563 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.642617 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:41Z","lastTransitionTime":"2025-10-02T09:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.745734 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.745808 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.745830 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.745861 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.745885 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:41Z","lastTransitionTime":"2025-10-02T09:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.848850 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.848902 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.848912 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.848926 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.848937 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:41Z","lastTransitionTime":"2025-10-02T09:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.912810 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.912857 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.912911 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.912843 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:41 crc kubenswrapper[4934]: E1002 09:49:41.912994 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:49:41 crc kubenswrapper[4934]: E1002 09:49:41.913074 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:49:41 crc kubenswrapper[4934]: E1002 09:49:41.913156 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:49:41 crc kubenswrapper[4934]: E1002 09:49:41.913202 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.952974 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.953020 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.953037 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.953057 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:41 crc kubenswrapper[4934]: I1002 09:49:41.953068 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:41Z","lastTransitionTime":"2025-10-02T09:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.056015 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.056052 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.056063 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.056082 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.056093 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:42Z","lastTransitionTime":"2025-10-02T09:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.159840 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.159897 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.159910 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.159933 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.159945 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:42Z","lastTransitionTime":"2025-10-02T09:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.262829 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.262873 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.262885 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.262913 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.262926 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:42Z","lastTransitionTime":"2025-10-02T09:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.366194 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.366261 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.366272 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.366291 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.366304 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:42Z","lastTransitionTime":"2025-10-02T09:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.469146 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.469232 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.469256 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.469288 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.469311 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:42Z","lastTransitionTime":"2025-10-02T09:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.572762 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.572837 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.572858 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.572880 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.572893 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:42Z","lastTransitionTime":"2025-10-02T09:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.676348 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.676414 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.676426 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.676446 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.676458 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:42Z","lastTransitionTime":"2025-10-02T09:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.779734 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.779809 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.779822 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.779848 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.779866 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:42Z","lastTransitionTime":"2025-10-02T09:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.883200 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.883245 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.883258 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.883282 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.883300 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:42Z","lastTransitionTime":"2025-10-02T09:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.985849 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.985896 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.985908 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.985929 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:42 crc kubenswrapper[4934]: I1002 09:49:42.985946 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:42Z","lastTransitionTime":"2025-10-02T09:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.089816 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.089898 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.089922 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.089954 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.089995 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:43Z","lastTransitionTime":"2025-10-02T09:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.193915 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.193962 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.193973 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.193991 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.194004 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:43Z","lastTransitionTime":"2025-10-02T09:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.296573 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.296649 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.296662 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.296682 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.296695 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:43Z","lastTransitionTime":"2025-10-02T09:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.400449 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.400516 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.400533 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.400561 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.400617 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:43Z","lastTransitionTime":"2025-10-02T09:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.503711 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.503796 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.503821 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.503856 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.503881 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:43Z","lastTransitionTime":"2025-10-02T09:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.606373 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.606412 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.606423 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.606439 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.606448 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:43Z","lastTransitionTime":"2025-10-02T09:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.709509 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.709567 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.709604 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.709624 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.709636 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:43Z","lastTransitionTime":"2025-10-02T09:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.812046 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.812113 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.812131 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.812156 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.812173 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:43Z","lastTransitionTime":"2025-10-02T09:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.912145 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.912181 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.912251 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:43 crc kubenswrapper[4934]: E1002 09:49:43.912516 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.912547 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:43 crc kubenswrapper[4934]: E1002 09:49:43.912749 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:49:43 crc kubenswrapper[4934]: E1002 09:49:43.912792 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:49:43 crc kubenswrapper[4934]: E1002 09:49:43.912877 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.914514 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.914542 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.914552 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.914616 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:43 crc kubenswrapper[4934]: I1002 09:49:43.914635 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:43Z","lastTransitionTime":"2025-10-02T09:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.017110 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.017156 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.017172 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.017230 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.017248 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:44Z","lastTransitionTime":"2025-10-02T09:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.119792 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.119872 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.119893 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.119916 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.119935 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:44Z","lastTransitionTime":"2025-10-02T09:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.223512 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.223603 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.223625 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.223653 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.223686 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:44Z","lastTransitionTime":"2025-10-02T09:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.326850 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.326908 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.326923 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.326946 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.326961 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:44Z","lastTransitionTime":"2025-10-02T09:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.430021 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.430126 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.430179 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.430214 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.430236 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:44Z","lastTransitionTime":"2025-10-02T09:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.533493 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.533681 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.533734 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.533929 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.533973 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:44Z","lastTransitionTime":"2025-10-02T09:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.636961 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.637006 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.637017 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.637034 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.637046 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:44Z","lastTransitionTime":"2025-10-02T09:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.739380 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.739430 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.739445 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.739464 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.739477 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:44Z","lastTransitionTime":"2025-10-02T09:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.842164 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.842209 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.842218 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.842235 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.842245 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:44Z","lastTransitionTime":"2025-10-02T09:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.945646 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.945692 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.945706 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.945722 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:44 crc kubenswrapper[4934]: I1002 09:49:44.945732 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:44Z","lastTransitionTime":"2025-10-02T09:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.049186 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.049230 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.049240 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.049258 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.049269 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:45Z","lastTransitionTime":"2025-10-02T09:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.152342 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.152403 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.152425 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.152452 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.152472 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:45Z","lastTransitionTime":"2025-10-02T09:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.255835 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.255916 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.255948 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.255976 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.255999 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:45Z","lastTransitionTime":"2025-10-02T09:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.358823 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.358879 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.358895 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.358919 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.358938 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:45Z","lastTransitionTime":"2025-10-02T09:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.389347 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.389403 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.389414 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.389435 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.389447 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:45Z","lastTransitionTime":"2025-10-02T09:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:45 crc kubenswrapper[4934]: E1002 09:49:45.406768 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.413421 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.413479 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.413489 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.413510 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.413521 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:45Z","lastTransitionTime":"2025-10-02T09:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:45 crc kubenswrapper[4934]: E1002 09:49:45.432443 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.438543 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.438620 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.438633 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.438651 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.438661 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:45Z","lastTransitionTime":"2025-10-02T09:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:45 crc kubenswrapper[4934]: E1002 09:49:45.458842 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.463190 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.463260 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.463288 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.463322 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.463341 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:45Z","lastTransitionTime":"2025-10-02T09:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:45 crc kubenswrapper[4934]: E1002 09:49:45.480975 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.485354 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.485408 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.485427 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.485454 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.485472 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:45Z","lastTransitionTime":"2025-10-02T09:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:45 crc kubenswrapper[4934]: E1002 09:49:45.504096 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:45 crc kubenswrapper[4934]: E1002 09:49:45.504225 4934 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.506980 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.507189 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.507265 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.507348 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.507420 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:45Z","lastTransitionTime":"2025-10-02T09:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.566982 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.581056 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.592403 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.608192 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.610639 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.610689 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.610700 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.610719 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.610735 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:45Z","lastTransitionTime":"2025-10-02T09:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.631080 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.646318 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7cgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7cgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.668653 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.684705 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.705698 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.713058 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.713091 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.713100 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.713113 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.713123 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:45Z","lastTransitionTime":"2025-10-02T09:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.722559 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.737498 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.752775 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27680347ccabc1c590a7ffc59feeb56ad6123a3ae83806a893ac9283d4c6b5ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.769058 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.778146 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.788151 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.808337 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bfd39d2fcb3f63e1b033c6db4503d8ce5466e7af8555d0890174db12cca9773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfd39d2fcb3f63e1b033c6db4503d8ce5466e7af8555d0890174db12cca9773\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:33Z\\\",\\\"message\\\":\\\"tions: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z]\\\\nI1002 09:49:33.730641 6587 services_controller.go:434] Service openshift-operator-lifecycle-manager/packageserver-service retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{packageserver-service openshift-operator-lifecycle-manager a60a1f74-c6ff-4c81-96ae-27ba9796ba61 5485 0 2025-02-23 05:23:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true] map[] [{operators.coreos.com/v1alpha1 ClusterServiceVersion packageserver bbc08db6-5ba4-4fc4-b49d-26331e1e728b 0xc0076f7bcd 0xc0076f7bce}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:5443,Protocol:TCP,Port:5443,TargetPort:{0 5443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: packageserver,},ClusterIP:10.217.4.153,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.153],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-th4v6_openshift-ovn-kubernetes(1d4841b5-0469-461e-875c-25b9fe848141)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.815895 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.815927 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.815937 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.815951 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.815960 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:45Z","lastTransitionTime":"2025-10-02T09:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.821420 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://106b1f23a09ef32bc00222f00777a5a8c211b61c6f3d9fc4b67d7708f8091bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fc3a7a93196f00272230d3c06af1b408d42a806a5e65fbb3cf131ddfb37077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ffb9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.840850 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.855526 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851e2f96cda63332af97a90a26088ecdd7fcf2d0498b53e5d542aa532b048e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:45Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.912443 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.912469 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.912513 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.912558 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:45 crc kubenswrapper[4934]: E1002 09:49:45.912785 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:49:45 crc kubenswrapper[4934]: E1002 09:49:45.912969 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:49:45 crc kubenswrapper[4934]: E1002 09:49:45.913134 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:49:45 crc kubenswrapper[4934]: E1002 09:49:45.913252 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.918192 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.918229 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.918241 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.918265 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:45 crc kubenswrapper[4934]: I1002 09:49:45.918290 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:45Z","lastTransitionTime":"2025-10-02T09:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.021803 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.021857 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.021879 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.021906 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.021928 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:46Z","lastTransitionTime":"2025-10-02T09:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.124943 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.125000 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.125010 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.125030 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.125041 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:46Z","lastTransitionTime":"2025-10-02T09:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.228110 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.228184 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.228206 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.228235 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.228260 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:46Z","lastTransitionTime":"2025-10-02T09:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.330798 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.330841 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.330853 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.330868 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.330878 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:46Z","lastTransitionTime":"2025-10-02T09:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.433935 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.433986 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.433999 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.434015 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.434029 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:46Z","lastTransitionTime":"2025-10-02T09:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.536354 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.536415 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.536429 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.536450 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.536462 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:46Z","lastTransitionTime":"2025-10-02T09:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.638528 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.638570 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.638596 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.638649 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.638660 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:46Z","lastTransitionTime":"2025-10-02T09:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.741482 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.741538 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.741550 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.741567 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.741597 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:46Z","lastTransitionTime":"2025-10-02T09:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.843416 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.843464 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.843476 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.843493 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.843505 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:46Z","lastTransitionTime":"2025-10-02T09:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.946752 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.946793 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.946807 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.946824 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:46 crc kubenswrapper[4934]: I1002 09:49:46.946836 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:46Z","lastTransitionTime":"2025-10-02T09:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.049046 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.049085 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.049096 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.049114 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.049125 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:47Z","lastTransitionTime":"2025-10-02T09:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.151678 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.151730 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.151743 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.151758 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.151770 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:47Z","lastTransitionTime":"2025-10-02T09:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.254044 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.254105 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.254114 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.254133 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.254144 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:47Z","lastTransitionTime":"2025-10-02T09:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.357418 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.357467 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.357477 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.357494 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.357523 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:47Z","lastTransitionTime":"2025-10-02T09:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.460366 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.460454 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.460483 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.460515 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.460539 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:47Z","lastTransitionTime":"2025-10-02T09:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.563003 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.563051 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.563063 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.563080 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.563092 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:47Z","lastTransitionTime":"2025-10-02T09:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.665966 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.666029 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.666050 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.666080 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.666104 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:47Z","lastTransitionTime":"2025-10-02T09:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.768923 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.768987 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.768997 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.769017 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.769033 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:47Z","lastTransitionTime":"2025-10-02T09:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.871925 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.871976 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.871987 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.872007 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.872018 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:47Z","lastTransitionTime":"2025-10-02T09:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.912941 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.912991 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.912974 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.912946 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:47 crc kubenswrapper[4934]: E1002 09:49:47.913129 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:49:47 crc kubenswrapper[4934]: E1002 09:49:47.913273 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:49:47 crc kubenswrapper[4934]: E1002 09:49:47.913422 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:49:47 crc kubenswrapper[4934]: E1002 09:49:47.913549 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.974312 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.974370 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.974386 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.974411 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:47 crc kubenswrapper[4934]: I1002 09:49:47.974428 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:47Z","lastTransitionTime":"2025-10-02T09:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.077098 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.077163 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.077174 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.077188 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.077196 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:48Z","lastTransitionTime":"2025-10-02T09:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.179827 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.179866 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.179877 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.179894 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.179905 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:48Z","lastTransitionTime":"2025-10-02T09:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.282872 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.282923 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.282935 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.282953 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.282962 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:48Z","lastTransitionTime":"2025-10-02T09:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.385514 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.385559 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.385567 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.385621 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.385634 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:48Z","lastTransitionTime":"2025-10-02T09:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.488927 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.489010 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.489023 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.489041 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.489056 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:48Z","lastTransitionTime":"2025-10-02T09:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.592166 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.592207 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.592219 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.592239 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.592251 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:48Z","lastTransitionTime":"2025-10-02T09:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.694733 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.694768 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.694778 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.694793 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.694802 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:48Z","lastTransitionTime":"2025-10-02T09:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.796831 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.796868 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.796880 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.796897 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.796909 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:48Z","lastTransitionTime":"2025-10-02T09:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.900400 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.900726 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.900734 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.900750 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.900759 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:48Z","lastTransitionTime":"2025-10-02T09:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.930950 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:48Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.946662 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:48Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.959680 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:48Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.972873 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7cgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7cgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:48Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.986672 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:48Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:48 crc kubenswrapper[4934]: I1002 09:49:48.998785 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:48Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.003821 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.003863 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.003876 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.003895 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.003907 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:49Z","lastTransitionTime":"2025-10-02T09:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.014511 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:49Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.026700 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7186037-ba99-4a2b-8a22-6417ffc44b80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407acc937940ce03f746be72752c5efa2ce7e416c39a80c2763befdec18daa56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f14dcea7cf5a10dfe10430bfd3d6e41710d33685f43cdc634206a17bbb3697f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9af6ae65b40a9786691a168f58a28aa4e1f1fdb401976046e4657bbbc0d44fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6711e8eee36c5d872a158ccedc38e3739b94e96e19564babf96bd6fef2dad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c6711e8eee36c5d872a158ccedc38e3739b94e96e19564babf96bd6fef2dad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:49Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.042147 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:49Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.061247 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:49Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.078256 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bfd39d2fcb3f63e1b033c6db4503d8ce5466e7af8555d0890174db12cca9773\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfd39d2fcb3f63e1b033c6db4503d8ce5466e7af8555d0890174db12cca9773\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:33Z\\\",\\\"message\\\":\\\"tions: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z]\\\\nI1002 09:49:33.730641 6587 services_controller.go:434] Service openshift-operator-lifecycle-manager/packageserver-service retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{packageserver-service openshift-operator-lifecycle-manager a60a1f74-c6ff-4c81-96ae-27ba9796ba61 5485 0 2025-02-23 05:23:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true] map[] [{operators.coreos.com/v1alpha1 ClusterServiceVersion packageserver bbc08db6-5ba4-4fc4-b49d-26331e1e728b 0xc0076f7bcd 0xc0076f7bce}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:5443,Protocol:TCP,Port:5443,TargetPort:{0 5443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: packageserver,},ClusterIP:10.217.4.153,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.153],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-th4v6_openshift-ovn-kubernetes(1d4841b5-0469-461e-875c-25b9fe848141)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:49Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.092261 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://106b1f23a09ef32bc00222f00777a5a8c211b61c6f3d9fc4b67d7708f8091bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fc3a7a93196f00272230d3c06af1b408d42a806a5e65fbb3cf131ddfb37077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ffb9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:49Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.103945 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27680347ccabc1c590a7ffc59feeb56ad6123a3ae83806a893ac9283d4c6b5ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:49Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.106151 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.106177 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.106185 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.106199 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.106208 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:49Z","lastTransitionTime":"2025-10-02T09:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.116855 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:49Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.125726 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:49Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.135889 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:49Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.156402 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:49Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.169637 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851e2f96cda63332af97a90a26088ecdd7fcf2d0498b53e5d542aa532b048e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:49Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.208751 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.208789 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.208805 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.208826 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.208841 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:49Z","lastTransitionTime":"2025-10-02T09:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.311990 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.312031 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.312044 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.312063 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.312075 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:49Z","lastTransitionTime":"2025-10-02T09:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.415788 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.415870 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.415893 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.415923 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.415941 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:49Z","lastTransitionTime":"2025-10-02T09:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.518772 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.518811 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.518820 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.518835 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.518846 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:49Z","lastTransitionTime":"2025-10-02T09:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.621016 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.621061 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.621073 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.621095 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.621109 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:49Z","lastTransitionTime":"2025-10-02T09:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.723079 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.723133 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.723147 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.723169 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.723181 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:49Z","lastTransitionTime":"2025-10-02T09:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.826264 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.826342 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.826357 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.826374 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.826407 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:49Z","lastTransitionTime":"2025-10-02T09:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.912981 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.913027 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:49 crc kubenswrapper[4934]: E1002 09:49:49.913157 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.913185 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.913222 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:49 crc kubenswrapper[4934]: E1002 09:49:49.913408 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:49:49 crc kubenswrapper[4934]: E1002 09:49:49.913568 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:49:49 crc kubenswrapper[4934]: E1002 09:49:49.913693 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.928786 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.928830 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.928842 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.928860 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:49 crc kubenswrapper[4934]: I1002 09:49:49.928872 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:49Z","lastTransitionTime":"2025-10-02T09:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.032302 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.032373 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.032386 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.032405 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.032418 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:50Z","lastTransitionTime":"2025-10-02T09:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.134351 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.134389 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.134399 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.134415 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.134426 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:50Z","lastTransitionTime":"2025-10-02T09:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.238711 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.238789 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.238806 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.238836 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.238856 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:50Z","lastTransitionTime":"2025-10-02T09:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.341718 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.341769 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.341788 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.341814 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.341831 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:50Z","lastTransitionTime":"2025-10-02T09:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.444321 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.444452 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.444476 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.444507 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.444527 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:50Z","lastTransitionTime":"2025-10-02T09:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.547819 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.547881 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.547898 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.547922 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.547939 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:50Z","lastTransitionTime":"2025-10-02T09:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.650919 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.650990 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.651013 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.651046 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.651063 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:50Z","lastTransitionTime":"2025-10-02T09:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.753476 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.753545 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.753562 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.753618 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.753637 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:50Z","lastTransitionTime":"2025-10-02T09:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.856060 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.856123 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.856138 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.856160 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.856178 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:50Z","lastTransitionTime":"2025-10-02T09:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.958880 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.958972 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.958985 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.959007 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:50 crc kubenswrapper[4934]: I1002 09:49:50.959020 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:50Z","lastTransitionTime":"2025-10-02T09:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.061874 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.061941 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.061959 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.061988 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.062011 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:51Z","lastTransitionTime":"2025-10-02T09:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.165967 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.166029 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.166043 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.166066 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.166079 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:51Z","lastTransitionTime":"2025-10-02T09:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.268537 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.268608 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.268621 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.268638 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.268647 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:51Z","lastTransitionTime":"2025-10-02T09:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.371494 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.371804 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.371817 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.371837 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.371853 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:51Z","lastTransitionTime":"2025-10-02T09:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.474801 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.474856 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.474869 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.474886 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.474895 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:51Z","lastTransitionTime":"2025-10-02T09:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.577532 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.577596 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.577613 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.577632 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.577644 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:51Z","lastTransitionTime":"2025-10-02T09:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.680478 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.680542 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.680558 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.680610 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.680628 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:51Z","lastTransitionTime":"2025-10-02T09:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.783608 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.783636 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.783644 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.783657 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.783665 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:51Z","lastTransitionTime":"2025-10-02T09:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.886557 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.886649 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.886668 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.886690 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.886705 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:51Z","lastTransitionTime":"2025-10-02T09:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.912711 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.912753 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.912753 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.912744 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:51 crc kubenswrapper[4934]: E1002 09:49:51.912906 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:49:51 crc kubenswrapper[4934]: E1002 09:49:51.913029 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:49:51 crc kubenswrapper[4934]: E1002 09:49:51.913200 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:49:51 crc kubenswrapper[4934]: E1002 09:49:51.913294 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.989407 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.989495 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.989511 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.989536 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:51 crc kubenswrapper[4934]: I1002 09:49:51.989553 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:51Z","lastTransitionTime":"2025-10-02T09:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.092459 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.092505 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.092518 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.092534 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.092545 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:52Z","lastTransitionTime":"2025-10-02T09:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.195113 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.195159 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.195170 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.195197 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.195210 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:52Z","lastTransitionTime":"2025-10-02T09:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.297909 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.297980 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.297997 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.298022 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.298041 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:52Z","lastTransitionTime":"2025-10-02T09:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.400685 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.400733 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.400744 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.400762 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.400775 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:52Z","lastTransitionTime":"2025-10-02T09:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.504532 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.504607 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.504625 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.504649 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.504666 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:52Z","lastTransitionTime":"2025-10-02T09:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.608888 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.608973 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.608992 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.609020 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.609037 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:52Z","lastTransitionTime":"2025-10-02T09:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.712137 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.712188 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.712237 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.712256 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.712267 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:52Z","lastTransitionTime":"2025-10-02T09:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.814210 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.814245 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.814254 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.814269 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.814278 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:52Z","lastTransitionTime":"2025-10-02T09:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.916769 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.916816 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.916827 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.916842 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:52 crc kubenswrapper[4934]: I1002 09:49:52.916853 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:52Z","lastTransitionTime":"2025-10-02T09:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.019435 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.019486 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.019501 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.019524 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.019544 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:53Z","lastTransitionTime":"2025-10-02T09:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.122708 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.122748 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.122756 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.122771 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.122781 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:53Z","lastTransitionTime":"2025-10-02T09:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.225731 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.225781 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.225793 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.225811 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.225823 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:53Z","lastTransitionTime":"2025-10-02T09:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.328307 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.328364 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.328376 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.328395 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.328406 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:53Z","lastTransitionTime":"2025-10-02T09:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.431920 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.431976 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.431992 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.432016 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.432029 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:53Z","lastTransitionTime":"2025-10-02T09:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.534083 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.534118 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.534127 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.534140 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.534150 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:53Z","lastTransitionTime":"2025-10-02T09:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.636485 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.636523 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.636534 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.636549 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.636559 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:53Z","lastTransitionTime":"2025-10-02T09:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.739155 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.739196 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.739209 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.739226 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.739237 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:53Z","lastTransitionTime":"2025-10-02T09:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.841846 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.841879 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.841889 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.841903 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.841912 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:53Z","lastTransitionTime":"2025-10-02T09:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.913102 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.913200 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:53 crc kubenswrapper[4934]: E1002 09:49:53.913272 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.913229 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.913191 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:53 crc kubenswrapper[4934]: E1002 09:49:53.913341 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:49:53 crc kubenswrapper[4934]: E1002 09:49:53.913374 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:49:53 crc kubenswrapper[4934]: E1002 09:49:53.913474 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.944150 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.944199 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.944212 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.944230 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:53 crc kubenswrapper[4934]: I1002 09:49:53.944243 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:53Z","lastTransitionTime":"2025-10-02T09:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.047074 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.047118 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.047131 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.047152 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.047165 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:54Z","lastTransitionTime":"2025-10-02T09:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.149705 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.149784 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.149797 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.149814 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.149826 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:54Z","lastTransitionTime":"2025-10-02T09:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.253223 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.253268 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.253281 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.253298 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.253334 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:54Z","lastTransitionTime":"2025-10-02T09:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.356450 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.356513 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.356531 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.356560 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.356605 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:54Z","lastTransitionTime":"2025-10-02T09:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.459120 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.459187 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.459205 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.459229 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.459245 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:54Z","lastTransitionTime":"2025-10-02T09:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.561393 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.561443 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.561455 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.561475 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.561488 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:54Z","lastTransitionTime":"2025-10-02T09:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.664092 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.664133 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.664143 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.664160 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.664174 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:54Z","lastTransitionTime":"2025-10-02T09:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.766516 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.766555 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.766565 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.766594 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.766604 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:54Z","lastTransitionTime":"2025-10-02T09:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.868928 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.868966 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.868976 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.869009 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.869020 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:54Z","lastTransitionTime":"2025-10-02T09:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.913291 4934 scope.go:117] "RemoveContainer" containerID="9bfd39d2fcb3f63e1b033c6db4503d8ce5466e7af8555d0890174db12cca9773" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.972274 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.972533 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.972542 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.972557 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:54 crc kubenswrapper[4934]: I1002 09:49:54.972567 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:54Z","lastTransitionTime":"2025-10-02T09:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.080837 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.080884 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.080895 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.080912 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.080924 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:55Z","lastTransitionTime":"2025-10-02T09:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.183407 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.183453 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.183464 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.183484 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.183494 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:55Z","lastTransitionTime":"2025-10-02T09:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.285617 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.285664 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.285678 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.285697 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.285709 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:55Z","lastTransitionTime":"2025-10-02T09:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.318852 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-th4v6_1d4841b5-0469-461e-875c-25b9fe848141/ovnkube-controller/2.log" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.321322 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" event={"ID":"1d4841b5-0469-461e-875c-25b9fe848141","Type":"ContainerStarted","Data":"358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe"} Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.321672 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.347315 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfd39d2fcb3f63e1b033c6db4503d8ce5466e7af8555d0890174db12cca9773\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:33Z\\\",\\\"message\\\":\\\"tions: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z]\\\\nI1002 09:49:33.730641 6587 services_controller.go:434] Service openshift-operator-lifecycle-manager/packageserver-service retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{packageserver-service openshift-operator-lifecycle-manager a60a1f74-c6ff-4c81-96ae-27ba9796ba61 5485 0 2025-02-23 05:23:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true] map[] [{operators.coreos.com/v1alpha1 ClusterServiceVersion packageserver bbc08db6-5ba4-4fc4-b49d-26331e1e728b 0xc0076f7bcd 0xc0076f7bce}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:5443,Protocol:TCP,Port:5443,TargetPort:{0 5443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: packageserver,},ClusterIP:10.217.4.153,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.153],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.367993 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://106b1f23a09ef32bc00222f00777a5a8c211b61c6f3d9fc4b67d7708f8091bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fc3a7a93196f00272230d3c06af1b408d42a806a5e65fbb3cf131ddfb37077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ffb9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.443919 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.443947 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.443957 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.443975 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.443992 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:55Z","lastTransitionTime":"2025-10-02T09:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.447864 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27680347ccabc1c590a7ffc59feeb56ad6123a3ae83806a893ac9283d4c6b5ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.462813 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.472321 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.483646 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.505092 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.519288 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851e2f96cda63332af97a90a26088ecdd7fcf2d0498b53e5d542aa532b048e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.531931 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.546171 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.546207 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.546217 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.546263 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.546276 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:55Z","lastTransitionTime":"2025-10-02T09:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.546945 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.561165 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.571882 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7cgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7cgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.583956 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.595943 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.608091 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.624127 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7186037-ba99-4a2b-8a22-6417ffc44b80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407acc937940ce03f746be72752c5efa2ce7e416c39a80c2763befdec18daa56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f14dcea7cf5a10dfe10430bfd3d6e41710d33685f43cdc634206a17bbb3697f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9af6ae65b40a9786691a168f58a28aa4e1f1fdb401976046e4657bbbc0d44fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6711e8eee36c5d872a158ccedc38e3739b94e96e19564babf96bd6fef2dad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c6711e8eee36c5d872a158ccedc38e3739b94e96e19564babf96bd6fef2dad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.634935 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.647729 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.648367 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.648403 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.648417 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.648433 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.648444 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:55Z","lastTransitionTime":"2025-10-02T09:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.750677 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.750715 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.750728 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.750744 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.750756 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:55Z","lastTransitionTime":"2025-10-02T09:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.809706 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.809753 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.809765 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.809783 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.809796 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:55Z","lastTransitionTime":"2025-10-02T09:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:55 crc kubenswrapper[4934]: E1002 09:49:55.826565 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.830243 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.830283 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.830293 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.830309 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.830320 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:55Z","lastTransitionTime":"2025-10-02T09:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:55 crc kubenswrapper[4934]: E1002 09:49:55.853622 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.857150 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.857177 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.857184 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.857198 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.857208 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:55Z","lastTransitionTime":"2025-10-02T09:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.861050 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-metrics-certs\") pod \"network-metrics-daemon-n7cgz\" (UID: \"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\") " pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:55 crc kubenswrapper[4934]: E1002 09:49:55.861234 4934 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:49:55 crc kubenswrapper[4934]: E1002 09:49:55.861311 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-metrics-certs podName:0cbef3ee-c66f-47fa-94d3-5ce9892c403f nodeName:}" failed. No retries permitted until 2025-10-02 09:50:27.861291137 +0000 UTC m=+99.613932759 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-metrics-certs") pod "network-metrics-daemon-n7cgz" (UID: "0cbef3ee-c66f-47fa-94d3-5ce9892c403f") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:49:55 crc kubenswrapper[4934]: E1002 09:49:55.868484 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.871697 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.871913 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.872052 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.872172 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.872410 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:55Z","lastTransitionTime":"2025-10-02T09:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:55 crc kubenswrapper[4934]: E1002 09:49:55.886666 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.891752 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.891784 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.891794 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.891812 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.891825 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:55Z","lastTransitionTime":"2025-10-02T09:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:55 crc kubenswrapper[4934]: E1002 09:49:55.905169 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:55 crc kubenswrapper[4934]: E1002 09:49:55.905610 4934 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.907438 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.907700 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.907835 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.907954 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.908086 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:55Z","lastTransitionTime":"2025-10-02T09:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.912752 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.912787 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.912749 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:55 crc kubenswrapper[4934]: E1002 09:49:55.912854 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:49:55 crc kubenswrapper[4934]: I1002 09:49:55.912898 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:55 crc kubenswrapper[4934]: E1002 09:49:55.912934 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:49:55 crc kubenswrapper[4934]: E1002 09:49:55.913053 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:49:55 crc kubenswrapper[4934]: E1002 09:49:55.913182 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.010858 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.011167 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.011256 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.011352 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.011438 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:56Z","lastTransitionTime":"2025-10-02T09:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.114567 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.114687 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.114705 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.114732 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.114750 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:56Z","lastTransitionTime":"2025-10-02T09:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.216569 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.216626 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.216639 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.216657 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.216670 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:56Z","lastTransitionTime":"2025-10-02T09:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.318733 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.318808 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.318821 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.318838 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.318850 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:56Z","lastTransitionTime":"2025-10-02T09:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.325869 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-th4v6_1d4841b5-0469-461e-875c-25b9fe848141/ovnkube-controller/3.log" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.326510 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-th4v6_1d4841b5-0469-461e-875c-25b9fe848141/ovnkube-controller/2.log" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.328806 4934 generic.go:334] "Generic (PLEG): container finished" podID="1d4841b5-0469-461e-875c-25b9fe848141" containerID="358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe" exitCode=1 Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.328842 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" event={"ID":"1d4841b5-0469-461e-875c-25b9fe848141","Type":"ContainerDied","Data":"358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe"} Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.328881 4934 scope.go:117] "RemoveContainer" containerID="9bfd39d2fcb3f63e1b033c6db4503d8ce5466e7af8555d0890174db12cca9773" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.330285 4934 scope.go:117] "RemoveContainer" containerID="358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe" Oct 02 09:49:56 crc kubenswrapper[4934]: E1002 09:49:56.330849 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-th4v6_openshift-ovn-kubernetes(1d4841b5-0469-461e-875c-25b9fe848141)\"" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" podUID="1d4841b5-0469-461e-875c-25b9fe848141" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.345137 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.358726 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.375138 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.385750 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7cgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7cgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.394809 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.406615 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.417541 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7186037-ba99-4a2b-8a22-6417ffc44b80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407acc937940ce03f746be72752c5efa2ce7e416c39a80c2763befdec18daa56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f14dcea7cf5a10dfe10430bfd3d6e41710d33685f43cdc634206a17bbb3697f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9af6ae65b40a9786691a168f58a28aa4e1f1fdb401976046e4657bbbc0d44fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6711e8eee36c5d872a158ccedc38e3739b94e96e19564babf96bd6fef2dad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c6711e8eee36c5d872a158ccedc38e3739b94e96e19564babf96bd6fef2dad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.420777 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.420808 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.420816 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.420830 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.420839 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:56Z","lastTransitionTime":"2025-10-02T09:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.430290 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.444569 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.457185 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.470230 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://106b1f23a09ef32bc00222f00777a5a8c211b61c6f3d9fc4b67d7708f8091bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fc3a7a93196f00272230d3c06af1b408d42a806a5e65fbb3cf131ddfb37077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ffb9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.484700 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27680347ccabc1c590a7ffc59feeb56ad6123a3ae83806a893ac9283d4c6b5ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.495555 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.504891 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.529682 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.529741 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.529754 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.529776 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.529799 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:56Z","lastTransitionTime":"2025-10-02T09:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.539774 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.558373 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9bfd39d2fcb3f63e1b033c6db4503d8ce5466e7af8555d0890174db12cca9773\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:33Z\\\",\\\"message\\\":\\\"tions: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:33Z is after 2025-08-24T17:21:41Z]\\\\nI1002 09:49:33.730641 6587 services_controller.go:434] Service openshift-operator-lifecycle-manager/packageserver-service retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{packageserver-service openshift-operator-lifecycle-manager a60a1f74-c6ff-4c81-96ae-27ba9796ba61 5485 0 2025-02-23 05:23:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true] map[] [{operators.coreos.com/v1alpha1 ClusterServiceVersion packageserver bbc08db6-5ba4-4fc4-b49d-26331e1e728b 0xc0076f7bcd 0xc0076f7bce}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:5443,Protocol:TCP,Port:5443,TargetPort:{0 5443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: packageserver,},ClusterIP:10.217.4.153,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.153],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"message\\\":\\\" EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:9154, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}, services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_UDP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"UDP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}}\\\\nI1002 09:49:55.844402 6884 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 09:49:55.844456 6884 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.575906 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.590912 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851e2f96cda63332af97a90a26088ecdd7fcf2d0498b53e5d542aa532b048e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.632821 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.632877 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.632888 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.632906 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.632918 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:56Z","lastTransitionTime":"2025-10-02T09:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.735853 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.735928 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.735951 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.735983 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.736007 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:56Z","lastTransitionTime":"2025-10-02T09:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.838551 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.838653 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.838677 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.838705 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.838727 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:56Z","lastTransitionTime":"2025-10-02T09:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.941217 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.941264 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.941274 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.941290 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:56 crc kubenswrapper[4934]: I1002 09:49:56.941301 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:56Z","lastTransitionTime":"2025-10-02T09:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.044072 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.044126 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.044140 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.044160 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.044173 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:57Z","lastTransitionTime":"2025-10-02T09:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.146378 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.146428 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.146442 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.146462 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.146491 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:57Z","lastTransitionTime":"2025-10-02T09:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.249082 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.249121 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.249132 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.249148 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.249159 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:57Z","lastTransitionTime":"2025-10-02T09:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.332809 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-th4v6_1d4841b5-0469-461e-875c-25b9fe848141/ovnkube-controller/3.log" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.336304 4934 scope.go:117] "RemoveContainer" containerID="358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe" Oct 02 09:49:57 crc kubenswrapper[4934]: E1002 09:49:57.336474 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-th4v6_openshift-ovn-kubernetes(1d4841b5-0469-461e-875c-25b9fe848141)\"" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" podUID="1d4841b5-0469-461e-875c-25b9fe848141" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.348078 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.352035 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.352071 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.352081 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.352096 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.352107 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:57Z","lastTransitionTime":"2025-10-02T09:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.361128 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.374384 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.385520 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.396176 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.408375 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7186037-ba99-4a2b-8a22-6417ffc44b80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407acc937940ce03f746be72752c5efa2ce7e416c39a80c2763befdec18daa56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f14dcea7cf5a10dfe10430bfd3d6e41710d33685f43cdc634206a17bbb3697f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9af6ae65b40a9786691a168f58a28aa4e1f1fdb401976046e4657bbbc0d44fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6711e8eee36c5d872a158ccedc38e3739b94e96e19564babf96bd6fef2dad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c6711e8eee36c5d872a158ccedc38e3739b94e96e19564babf96bd6fef2dad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.421428 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.433782 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.454009 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"message\\\":\\\" EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:9154, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}, services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_UDP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"UDP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}}\\\\nI1002 09:49:55.844402 6884 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 09:49:55.844456 6884 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-th4v6_openshift-ovn-kubernetes(1d4841b5-0469-461e-875c-25b9fe848141)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.454946 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.454996 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.455013 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.455038 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.455056 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:57Z","lastTransitionTime":"2025-10-02T09:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.465548 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://106b1f23a09ef32bc00222f00777a5a8c211b61c6f3d9fc4b67d7708f8091bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fc3a7a93196f00272230d3c06af1b408d42a806a5e65fbb3cf131ddfb37077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ffb9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.479848 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27680347ccabc1c590a7ffc59feeb56ad6123a3ae83806a893ac9283d4c6b5ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.491357 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.503365 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851e2f96cda63332af97a90a26088ecdd7fcf2d0498b53e5d542aa532b048e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.525089 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.542283 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.555366 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7cgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7cgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.557075 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.557107 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.557116 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.557131 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.557141 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:57Z","lastTransitionTime":"2025-10-02T09:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.569148 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.582605 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.659533 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.659566 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.659600 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.659618 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.659627 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:57Z","lastTransitionTime":"2025-10-02T09:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.761089 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.761118 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.761126 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.761139 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.761148 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:57Z","lastTransitionTime":"2025-10-02T09:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.863914 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.863962 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.863974 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.863989 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.864000 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:57Z","lastTransitionTime":"2025-10-02T09:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.912765 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.912808 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.912853 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:57 crc kubenswrapper[4934]: E1002 09:49:57.912893 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.912948 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:57 crc kubenswrapper[4934]: E1002 09:49:57.913022 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:49:57 crc kubenswrapper[4934]: E1002 09:49:57.913055 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:49:57 crc kubenswrapper[4934]: E1002 09:49:57.913108 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.965765 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.965817 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.965833 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.965854 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:57 crc kubenswrapper[4934]: I1002 09:49:57.965869 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:57Z","lastTransitionTime":"2025-10-02T09:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.068176 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.068221 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.068235 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.068254 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.068266 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:58Z","lastTransitionTime":"2025-10-02T09:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.170741 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.170789 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.170801 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.170835 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.170848 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:58Z","lastTransitionTime":"2025-10-02T09:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.273296 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.273330 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.273341 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.273361 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.273376 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:58Z","lastTransitionTime":"2025-10-02T09:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.340098 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-79fxg_73968247-67dd-48cc-88a1-64afac657412/kube-multus/0.log" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.340180 4934 generic.go:334] "Generic (PLEG): container finished" podID="73968247-67dd-48cc-88a1-64afac657412" containerID="08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65" exitCode=1 Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.340231 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-79fxg" event={"ID":"73968247-67dd-48cc-88a1-64afac657412","Type":"ContainerDied","Data":"08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65"} Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.340994 4934 scope.go:117] "RemoveContainer" containerID="08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.354459 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.366751 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:57Z\\\",\\\"message\\\":\\\"2025-10-02T09:49:12+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_730f868f-8db1-47c0-9b6f-91623ab15850\\\\n2025-10-02T09:49:12+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_730f868f-8db1-47c0-9b6f-91623ab15850 to /host/opt/cni/bin/\\\\n2025-10-02T09:49:12Z [verbose] multus-daemon started\\\\n2025-10-02T09:49:12Z [verbose] Readiness Indicator file check\\\\n2025-10-02T09:49:57Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.375458 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.375523 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.375546 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.375607 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.375625 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:58Z","lastTransitionTime":"2025-10-02T09:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.378398 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7cgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7cgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.396005 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.408967 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7186037-ba99-4a2b-8a22-6417ffc44b80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407acc937940ce03f746be72752c5efa2ce7e416c39a80c2763befdec18daa56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f14dcea7cf5a10dfe10430bfd3d6e41710d33685f43cdc634206a17bbb3697f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9af6ae65b40a9786691a168f58a28aa4e1f1fdb401976046e4657bbbc0d44fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6711e8eee36c5d872a158ccedc38e3739b94e96e19564babf96bd6fef2dad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c6711e8eee36c5d872a158ccedc38e3739b94e96e19564babf96bd6fef2dad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.420256 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.431124 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.444063 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.455309 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.472266 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.477784 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.477819 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.477828 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.477843 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.477853 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:58Z","lastTransitionTime":"2025-10-02T09:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.484999 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.496702 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.507296 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.527807 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"message\\\":\\\" EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:9154, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}, services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_UDP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"UDP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}}\\\\nI1002 09:49:55.844402 6884 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 09:49:55.844456 6884 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-th4v6_openshift-ovn-kubernetes(1d4841b5-0469-461e-875c-25b9fe848141)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.539981 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://106b1f23a09ef32bc00222f00777a5a8c211b61c6f3d9fc4b67d7708f8091bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fc3a7a93196f00272230d3c06af1b408d42a806a5e65fbb3cf131ddfb37077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ffb9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.553594 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27680347ccabc1c590a7ffc59feeb56ad6123a3ae83806a893ac9283d4c6b5ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.577608 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.580008 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.580043 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.580054 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.580071 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.580082 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:58Z","lastTransitionTime":"2025-10-02T09:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.592988 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851e2f96cda63332af97a90a26088ecdd7fcf2d0498b53e5d542aa532b048e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.682008 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.682043 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.682055 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.682073 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.682085 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:58Z","lastTransitionTime":"2025-10-02T09:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.784529 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.784586 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.784597 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.784613 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.784624 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:58Z","lastTransitionTime":"2025-10-02T09:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.887469 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.887506 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.887517 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.887535 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.887547 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:58Z","lastTransitionTime":"2025-10-02T09:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.931714 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.944121 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7186037-ba99-4a2b-8a22-6417ffc44b80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407acc937940ce03f746be72752c5efa2ce7e416c39a80c2763befdec18daa56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f14dcea7cf5a10dfe10430bfd3d6e41710d33685f43cdc634206a17bbb3697f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9af6ae65b40a9786691a168f58a28aa4e1f1fdb401976046e4657bbbc0d44fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6711e8eee36c5d872a158ccedc38e3739b94e96e19564babf96bd6fef2dad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c6711e8eee36c5d872a158ccedc38e3739b94e96e19564babf96bd6fef2dad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.958768 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.970914 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.987788 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.989415 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.989449 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.989459 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.989477 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.989487 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:58Z","lastTransitionTime":"2025-10-02T09:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:58 crc kubenswrapper[4934]: I1002 09:49:58.997813 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:58Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.011099 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27680347ccabc1c590a7ffc59feeb56ad6123a3ae83806a893ac9283d4c6b5ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.022314 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.032522 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.044674 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.063057 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"message\\\":\\\" EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:9154, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}, services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_UDP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"UDP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}}\\\\nI1002 09:49:55.844402 6884 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 09:49:55.844456 6884 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-th4v6_openshift-ovn-kubernetes(1d4841b5-0469-461e-875c-25b9fe848141)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.075833 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://106b1f23a09ef32bc00222f00777a5a8c211b61c6f3d9fc4b67d7708f8091bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fc3a7a93196f00272230d3c06af1b408d42a806a5e65fbb3cf131ddfb37077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ffb9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.091910 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.091954 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.091967 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.091984 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.091996 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:59Z","lastTransitionTime":"2025-10-02T09:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.103147 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.116883 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851e2f96cda63332af97a90a26088ecdd7fcf2d0498b53e5d542aa532b048e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.129035 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.139175 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.149655 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:57Z\\\",\\\"message\\\":\\\"2025-10-02T09:49:12+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_730f868f-8db1-47c0-9b6f-91623ab15850\\\\n2025-10-02T09:49:12+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_730f868f-8db1-47c0-9b6f-91623ab15850 to /host/opt/cni/bin/\\\\n2025-10-02T09:49:12Z [verbose] multus-daemon started\\\\n2025-10-02T09:49:12Z [verbose] Readiness Indicator file check\\\\n2025-10-02T09:49:57Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.159353 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7cgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7cgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.194464 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.194510 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.194523 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.194539 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.194550 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:59Z","lastTransitionTime":"2025-10-02T09:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.296338 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.296371 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.296379 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.296394 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.296404 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:59Z","lastTransitionTime":"2025-10-02T09:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.344726 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-79fxg_73968247-67dd-48cc-88a1-64afac657412/kube-multus/0.log" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.344788 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-79fxg" event={"ID":"73968247-67dd-48cc-88a1-64afac657412","Type":"ContainerStarted","Data":"84da7560cede71d06bd39eeee26703d00479335105cce31a29648fbd36512917"} Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.364746 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27680347ccabc1c590a7ffc59feeb56ad6123a3ae83806a893ac9283d4c6b5ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.376615 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.388051 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.398189 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.398962 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.398991 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.399001 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.399016 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.399026 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:59Z","lastTransitionTime":"2025-10-02T09:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.417544 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"message\\\":\\\" EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:9154, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}, services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_UDP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"UDP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}}\\\\nI1002 09:49:55.844402 6884 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 09:49:55.844456 6884 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-th4v6_openshift-ovn-kubernetes(1d4841b5-0469-461e-875c-25b9fe848141)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.431107 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://106b1f23a09ef32bc00222f00777a5a8c211b61c6f3d9fc4b67d7708f8091bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fc3a7a93196f00272230d3c06af1b408d42a806a5e65fbb3cf131ddfb37077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ffb9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.451241 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.464351 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851e2f96cda63332af97a90a26088ecdd7fcf2d0498b53e5d542aa532b048e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.477085 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.489389 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.501039 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84da7560cede71d06bd39eeee26703d00479335105cce31a29648fbd36512917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:57Z\\\",\\\"message\\\":\\\"2025-10-02T09:49:12+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_730f868f-8db1-47c0-9b6f-91623ab15850\\\\n2025-10-02T09:49:12+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_730f868f-8db1-47c0-9b6f-91623ab15850 to /host/opt/cni/bin/\\\\n2025-10-02T09:49:12Z [verbose] multus-daemon started\\\\n2025-10-02T09:49:12Z [verbose] Readiness Indicator file check\\\\n2025-10-02T09:49:57Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.501725 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.501795 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.501808 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.501823 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.501834 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:59Z","lastTransitionTime":"2025-10-02T09:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.510594 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7cgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7cgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.521473 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.532291 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7186037-ba99-4a2b-8a22-6417ffc44b80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407acc937940ce03f746be72752c5efa2ce7e416c39a80c2763befdec18daa56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f14dcea7cf5a10dfe10430bfd3d6e41710d33685f43cdc634206a17bbb3697f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9af6ae65b40a9786691a168f58a28aa4e1f1fdb401976046e4657bbbc0d44fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6711e8eee36c5d872a158ccedc38e3739b94e96e19564babf96bd6fef2dad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c6711e8eee36c5d872a158ccedc38e3739b94e96e19564babf96bd6fef2dad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.543859 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.555926 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.567297 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.577450 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:49:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.604412 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.604448 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.604459 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.604476 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.604488 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:59Z","lastTransitionTime":"2025-10-02T09:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.707190 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.707476 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.707560 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.707681 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.707780 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:59Z","lastTransitionTime":"2025-10-02T09:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.809552 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.809838 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.809916 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.810002 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.810085 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:59Z","lastTransitionTime":"2025-10-02T09:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.912458 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.912458 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.912490 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.912622 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:49:59 crc kubenswrapper[4934]: E1002 09:49:59.912712 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:49:59 crc kubenswrapper[4934]: E1002 09:49:59.912822 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.912912 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.912956 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:49:59 crc kubenswrapper[4934]: E1002 09:49:59.912930 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.912969 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:49:59 crc kubenswrapper[4934]: E1002 09:49:59.912976 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.913067 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:49:59 crc kubenswrapper[4934]: I1002 09:49:59.913085 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:49:59Z","lastTransitionTime":"2025-10-02T09:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.015695 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.015744 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.015758 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.015774 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.015786 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:00Z","lastTransitionTime":"2025-10-02T09:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.117764 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.117818 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.117835 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.117857 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.117873 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:00Z","lastTransitionTime":"2025-10-02T09:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.220476 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.220515 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.220524 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.220538 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.220547 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:00Z","lastTransitionTime":"2025-10-02T09:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.323105 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.323146 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.323158 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.323174 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.323186 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:00Z","lastTransitionTime":"2025-10-02T09:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.425690 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.425741 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.425755 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.425776 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.425790 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:00Z","lastTransitionTime":"2025-10-02T09:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.527801 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.527845 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.527854 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.527870 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.527878 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:00Z","lastTransitionTime":"2025-10-02T09:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.630347 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.630399 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.630411 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.630433 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.630445 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:00Z","lastTransitionTime":"2025-10-02T09:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.732811 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.732860 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.732871 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.732887 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.732901 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:00Z","lastTransitionTime":"2025-10-02T09:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.835838 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.835881 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.835897 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.835921 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.835937 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:00Z","lastTransitionTime":"2025-10-02T09:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.938520 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.938559 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.938570 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.938620 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:00 crc kubenswrapper[4934]: I1002 09:50:00.938632 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:00Z","lastTransitionTime":"2025-10-02T09:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.040649 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.040693 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.040705 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.040726 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.040738 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:01Z","lastTransitionTime":"2025-10-02T09:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.146016 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.146217 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.146237 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.146255 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.146271 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:01Z","lastTransitionTime":"2025-10-02T09:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.248453 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.248478 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.248486 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.248500 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.248510 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:01Z","lastTransitionTime":"2025-10-02T09:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.350893 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.350916 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.350924 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.350938 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.350947 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:01Z","lastTransitionTime":"2025-10-02T09:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.453393 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.453437 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.453449 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.453464 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.453474 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:01Z","lastTransitionTime":"2025-10-02T09:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.556335 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.556375 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.556385 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.556401 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.556413 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:01Z","lastTransitionTime":"2025-10-02T09:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.659309 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.659370 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.659385 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.659407 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.659422 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:01Z","lastTransitionTime":"2025-10-02T09:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.762095 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.762135 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.762159 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.762195 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.762211 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:01Z","lastTransitionTime":"2025-10-02T09:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.865978 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.866062 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.866085 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.866115 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.866194 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:01Z","lastTransitionTime":"2025-10-02T09:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.912806 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.912845 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.912953 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:50:01 crc kubenswrapper[4934]: E1002 09:50:01.913110 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.913158 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:50:01 crc kubenswrapper[4934]: E1002 09:50:01.913291 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:50:01 crc kubenswrapper[4934]: E1002 09:50:01.913466 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:50:01 crc kubenswrapper[4934]: E1002 09:50:01.913685 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.970096 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.970185 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.970196 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.970211 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:01 crc kubenswrapper[4934]: I1002 09:50:01.970223 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:01Z","lastTransitionTime":"2025-10-02T09:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.074319 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.074621 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.074712 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.075030 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.075117 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:02Z","lastTransitionTime":"2025-10-02T09:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.178184 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.178228 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.178237 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.178253 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.178264 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:02Z","lastTransitionTime":"2025-10-02T09:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.280542 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.280591 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.280602 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.280619 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.280631 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:02Z","lastTransitionTime":"2025-10-02T09:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.382955 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.383039 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.383054 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.383081 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.383100 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:02Z","lastTransitionTime":"2025-10-02T09:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.485682 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.485721 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.485737 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.485754 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.485765 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:02Z","lastTransitionTime":"2025-10-02T09:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.588432 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.588548 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.588616 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.588638 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.588673 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:02Z","lastTransitionTime":"2025-10-02T09:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.691091 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.691546 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.691569 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.691655 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.691680 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:02Z","lastTransitionTime":"2025-10-02T09:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.793362 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.793396 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.793410 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.793428 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.793437 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:02Z","lastTransitionTime":"2025-10-02T09:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.896418 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.896460 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.896468 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.896482 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.896491 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:02Z","lastTransitionTime":"2025-10-02T09:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.999070 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.999111 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.999120 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.999136 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:02 crc kubenswrapper[4934]: I1002 09:50:02.999146 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:02Z","lastTransitionTime":"2025-10-02T09:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.103730 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.103766 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.103776 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.103796 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.103807 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:03Z","lastTransitionTime":"2025-10-02T09:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.206197 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.206251 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.206262 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.206280 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.206292 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:03Z","lastTransitionTime":"2025-10-02T09:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.309332 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.309380 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.309392 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.309413 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.309428 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:03Z","lastTransitionTime":"2025-10-02T09:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.412211 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.412294 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.412317 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.412335 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.412347 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:03Z","lastTransitionTime":"2025-10-02T09:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.514414 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.514459 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.514468 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.514483 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.514492 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:03Z","lastTransitionTime":"2025-10-02T09:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.616900 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.616951 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.616963 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.616976 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.616985 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:03Z","lastTransitionTime":"2025-10-02T09:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.719706 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.719761 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.719777 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.719799 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.719814 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:03Z","lastTransitionTime":"2025-10-02T09:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.823197 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.823258 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.823282 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.823309 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.823324 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:03Z","lastTransitionTime":"2025-10-02T09:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.912520 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:50:03 crc kubenswrapper[4934]: E1002 09:50:03.912653 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.912666 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:50:03 crc kubenswrapper[4934]: E1002 09:50:03.912841 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.912869 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.912943 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:50:03 crc kubenswrapper[4934]: E1002 09:50:03.913095 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:50:03 crc kubenswrapper[4934]: E1002 09:50:03.913229 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.926027 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.926055 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.926064 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.926076 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:03 crc kubenswrapper[4934]: I1002 09:50:03.926086 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:03Z","lastTransitionTime":"2025-10-02T09:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.028620 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.028657 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.028666 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.028679 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.028687 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:04Z","lastTransitionTime":"2025-10-02T09:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.130799 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.130826 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.130837 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.130850 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.130858 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:04Z","lastTransitionTime":"2025-10-02T09:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.233335 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.233408 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.233421 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.233443 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.233480 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:04Z","lastTransitionTime":"2025-10-02T09:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.335679 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.335992 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.336103 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.336184 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.336281 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:04Z","lastTransitionTime":"2025-10-02T09:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.439061 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.439101 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.439112 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.439129 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.439140 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:04Z","lastTransitionTime":"2025-10-02T09:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.541839 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.542086 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.542095 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.542111 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.542120 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:04Z","lastTransitionTime":"2025-10-02T09:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.644715 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.645014 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.645105 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.645187 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.645260 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:04Z","lastTransitionTime":"2025-10-02T09:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.748068 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.748147 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.748166 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.748193 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.748210 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:04Z","lastTransitionTime":"2025-10-02T09:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.851078 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.851121 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.851133 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.851150 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.851161 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:04Z","lastTransitionTime":"2025-10-02T09:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.953398 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.953448 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.953461 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.953479 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:04 crc kubenswrapper[4934]: I1002 09:50:04.953492 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:04Z","lastTransitionTime":"2025-10-02T09:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.056382 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.056448 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.056506 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.056535 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.056560 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:05Z","lastTransitionTime":"2025-10-02T09:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.159365 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.159407 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.159419 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.159436 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.159450 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:05Z","lastTransitionTime":"2025-10-02T09:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.261764 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.261807 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.261819 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.261837 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.261851 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:05Z","lastTransitionTime":"2025-10-02T09:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.364001 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.364066 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.364081 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.364099 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.364117 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:05Z","lastTransitionTime":"2025-10-02T09:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.467070 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.467135 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.467155 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.467178 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.467194 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:05Z","lastTransitionTime":"2025-10-02T09:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.569006 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.569073 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.569091 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.569114 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.569131 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:05Z","lastTransitionTime":"2025-10-02T09:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.671443 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.671477 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.671485 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.671501 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.671511 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:05Z","lastTransitionTime":"2025-10-02T09:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.773642 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.773691 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.773705 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.773721 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.773733 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:05Z","lastTransitionTime":"2025-10-02T09:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.876267 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.876326 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.876349 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.876377 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.876399 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:05Z","lastTransitionTime":"2025-10-02T09:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.912757 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.912751 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:50:05 crc kubenswrapper[4934]: E1002 09:50:05.912892 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.912938 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.912992 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:50:05 crc kubenswrapper[4934]: E1002 09:50:05.913126 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:50:05 crc kubenswrapper[4934]: E1002 09:50:05.913211 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:50:05 crc kubenswrapper[4934]: E1002 09:50:05.913291 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.979515 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.979601 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.979618 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.979644 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:05 crc kubenswrapper[4934]: I1002 09:50:05.979660 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:05Z","lastTransitionTime":"2025-10-02T09:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.083118 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.083459 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.083532 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.083658 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.083748 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:06Z","lastTransitionTime":"2025-10-02T09:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.084985 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.085040 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.085074 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.085122 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.085132 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:06Z","lastTransitionTime":"2025-10-02T09:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:06 crc kubenswrapper[4934]: E1002 09:50:06.101161 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.105074 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.105143 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.105154 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.105172 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.105183 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:06Z","lastTransitionTime":"2025-10-02T09:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:06 crc kubenswrapper[4934]: E1002 09:50:06.117369 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.120529 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.120555 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.120564 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.120591 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.120600 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:06Z","lastTransitionTime":"2025-10-02T09:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:06 crc kubenswrapper[4934]: E1002 09:50:06.130548 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.133438 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.133465 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.133476 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.133548 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.133566 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:06Z","lastTransitionTime":"2025-10-02T09:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:06 crc kubenswrapper[4934]: E1002 09:50:06.143318 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.146690 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.146722 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.146733 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.146749 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.146760 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:06Z","lastTransitionTime":"2025-10-02T09:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:06 crc kubenswrapper[4934]: E1002 09:50:06.157310 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:06 crc kubenswrapper[4934]: E1002 09:50:06.157469 4934 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.186337 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.186370 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.186380 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.186394 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.186405 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:06Z","lastTransitionTime":"2025-10-02T09:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.288383 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.288417 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.288427 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.288444 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.288454 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:06Z","lastTransitionTime":"2025-10-02T09:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.390468 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.390508 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.390519 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.390535 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.390544 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:06Z","lastTransitionTime":"2025-10-02T09:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.495966 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.496004 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.496016 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.496036 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.496047 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:06Z","lastTransitionTime":"2025-10-02T09:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.598605 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.598651 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.598674 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.598703 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.598721 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:06Z","lastTransitionTime":"2025-10-02T09:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.700972 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.701061 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.701073 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.701095 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.701107 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:06Z","lastTransitionTime":"2025-10-02T09:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.804264 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.804297 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.804307 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.804321 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.804329 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:06Z","lastTransitionTime":"2025-10-02T09:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.906723 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.906762 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.906774 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.906789 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:06 crc kubenswrapper[4934]: I1002 09:50:06.906801 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:06Z","lastTransitionTime":"2025-10-02T09:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.010603 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.010649 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.010659 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.010674 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.010683 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:07Z","lastTransitionTime":"2025-10-02T09:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.117009 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.117125 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.117180 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.117214 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.117229 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:07Z","lastTransitionTime":"2025-10-02T09:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.220121 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.220171 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.220182 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.220201 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.220213 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:07Z","lastTransitionTime":"2025-10-02T09:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.322929 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.322967 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.322980 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.322996 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.323007 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:07Z","lastTransitionTime":"2025-10-02T09:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.425460 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.425502 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.425514 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.425534 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.425545 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:07Z","lastTransitionTime":"2025-10-02T09:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.528226 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.528278 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.528294 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.528317 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.528330 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:07Z","lastTransitionTime":"2025-10-02T09:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.630816 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.631660 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.631820 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.631965 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.632103 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:07Z","lastTransitionTime":"2025-10-02T09:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.734694 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.734761 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.734780 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.734804 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.734820 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:07Z","lastTransitionTime":"2025-10-02T09:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.837954 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.838017 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.838038 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.838068 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.838088 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:07Z","lastTransitionTime":"2025-10-02T09:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.912917 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.912975 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:50:07 crc kubenswrapper[4934]: E1002 09:50:07.913047 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.913223 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:50:07 crc kubenswrapper[4934]: E1002 09:50:07.913257 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:50:07 crc kubenswrapper[4934]: E1002 09:50:07.913435 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.913909 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:50:07 crc kubenswrapper[4934]: E1002 09:50:07.914257 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.940900 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.940956 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.940970 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.940991 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:07 crc kubenswrapper[4934]: I1002 09:50:07.941010 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:07Z","lastTransitionTime":"2025-10-02T09:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.044132 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.044429 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.044535 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.044727 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.044868 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:08Z","lastTransitionTime":"2025-10-02T09:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.147353 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.147418 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.147439 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.147467 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.147488 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:08Z","lastTransitionTime":"2025-10-02T09:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.250558 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.250621 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.250633 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.250650 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.250663 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:08Z","lastTransitionTime":"2025-10-02T09:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.353087 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.353121 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.353133 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.353150 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.353163 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:08Z","lastTransitionTime":"2025-10-02T09:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.455666 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.455724 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.455740 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.455763 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.455781 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:08Z","lastTransitionTime":"2025-10-02T09:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.558294 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.558322 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.558331 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.558344 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.558352 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:08Z","lastTransitionTime":"2025-10-02T09:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.660990 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.661046 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.661058 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.661076 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.661088 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:08Z","lastTransitionTime":"2025-10-02T09:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.763062 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.763099 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.763109 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.763123 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.763136 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:08Z","lastTransitionTime":"2025-10-02T09:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.866013 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.866067 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.866083 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.866107 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.866125 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:08Z","lastTransitionTime":"2025-10-02T09:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.925482 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.939464 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.970466 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"message\\\":\\\" EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:9154, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}, services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_UDP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"UDP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}}\\\\nI1002 09:49:55.844402 6884 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 09:49:55.844456 6884 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-th4v6_openshift-ovn-kubernetes(1d4841b5-0469-461e-875c-25b9fe848141)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.971112 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.971143 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.971152 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.971167 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.971176 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:08Z","lastTransitionTime":"2025-10-02T09:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:08 crc kubenswrapper[4934]: I1002 09:50:08.984516 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://106b1f23a09ef32bc00222f00777a5a8c211b61c6f3d9fc4b67d7708f8091bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fc3a7a93196f00272230d3c06af1b408d42a806a5e65fbb3cf131ddfb37077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ffb9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.002991 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27680347ccabc1c590a7ffc59feeb56ad6123a3ae83806a893ac9283d4c6b5ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:09Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.018296 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:09Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.034521 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851e2f96cda63332af97a90a26088ecdd7fcf2d0498b53e5d542aa532b048e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:09Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.057831 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:09Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.070491 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84da7560cede71d06bd39eeee26703d00479335105cce31a29648fbd36512917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:57Z\\\",\\\"message\\\":\\\"2025-10-02T09:49:12+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_730f868f-8db1-47c0-9b6f-91623ab15850\\\\n2025-10-02T09:49:12+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_730f868f-8db1-47c0-9b6f-91623ab15850 to /host/opt/cni/bin/\\\\n2025-10-02T09:49:12Z [verbose] multus-daemon started\\\\n2025-10-02T09:49:12Z [verbose] Readiness Indicator file check\\\\n2025-10-02T09:49:57Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:09Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.073329 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.073380 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.073394 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.073412 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.073423 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:09Z","lastTransitionTime":"2025-10-02T09:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.081466 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7cgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7cgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:09Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.093865 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:09Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.105205 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:09Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.120309 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:09Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.132942 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:09Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.143881 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:09Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.154320 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:09Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.167375 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:09Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.175739 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.175765 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.175773 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.175786 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.175794 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:09Z","lastTransitionTime":"2025-10-02T09:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.178044 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7186037-ba99-4a2b-8a22-6417ffc44b80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407acc937940ce03f746be72752c5efa2ce7e416c39a80c2763befdec18daa56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f14dcea7cf5a10dfe10430bfd3d6e41710d33685f43cdc634206a17bbb3697f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9af6ae65b40a9786691a168f58a28aa4e1f1fdb401976046e4657bbbc0d44fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6711e8eee36c5d872a158ccedc38e3739b94e96e19564babf96bd6fef2dad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c6711e8eee36c5d872a158ccedc38e3739b94e96e19564babf96bd6fef2dad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:09Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.279242 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.279299 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.279310 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.279329 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.279340 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:09Z","lastTransitionTime":"2025-10-02T09:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.380932 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.380973 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.380983 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.380997 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.381006 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:09Z","lastTransitionTime":"2025-10-02T09:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.483909 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.483945 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.483955 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.483970 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.483981 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:09Z","lastTransitionTime":"2025-10-02T09:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.586223 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.586252 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.586260 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.586273 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.586281 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:09Z","lastTransitionTime":"2025-10-02T09:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.689466 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.689523 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.689541 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.689565 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.689614 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:09Z","lastTransitionTime":"2025-10-02T09:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.792417 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.792479 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.792496 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.792519 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.792535 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:09Z","lastTransitionTime":"2025-10-02T09:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.895287 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.895367 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.895429 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.895451 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.895463 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:09Z","lastTransitionTime":"2025-10-02T09:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.913007 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:50:09 crc kubenswrapper[4934]: E1002 09:50:09.913173 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.913202 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.913218 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.913243 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:50:09 crc kubenswrapper[4934]: E1002 09:50:09.913320 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:50:09 crc kubenswrapper[4934]: E1002 09:50:09.913397 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:50:09 crc kubenswrapper[4934]: E1002 09:50:09.913451 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.998223 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.998262 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.998272 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.998286 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:09 crc kubenswrapper[4934]: I1002 09:50:09.998295 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:09Z","lastTransitionTime":"2025-10-02T09:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.100734 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.100786 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.100795 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.100814 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.100823 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:10Z","lastTransitionTime":"2025-10-02T09:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.203879 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.203915 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.203925 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.203941 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.203952 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:10Z","lastTransitionTime":"2025-10-02T09:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.305699 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.305743 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.305751 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.305766 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.305775 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:10Z","lastTransitionTime":"2025-10-02T09:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.408379 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.408410 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.408420 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.408435 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.408446 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:10Z","lastTransitionTime":"2025-10-02T09:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.511291 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.511353 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.511375 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.511405 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.511429 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:10Z","lastTransitionTime":"2025-10-02T09:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.614272 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.614311 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.614321 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.614335 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.614344 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:10Z","lastTransitionTime":"2025-10-02T09:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.716786 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.716853 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.716890 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.716918 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.716936 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:10Z","lastTransitionTime":"2025-10-02T09:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.819570 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.819734 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.819762 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.819790 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.819810 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:10Z","lastTransitionTime":"2025-10-02T09:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.922209 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.922256 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.922272 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.922348 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:10 crc kubenswrapper[4934]: I1002 09:50:10.922365 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:10Z","lastTransitionTime":"2025-10-02T09:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.029033 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.029081 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.029091 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.029108 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.029118 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:11Z","lastTransitionTime":"2025-10-02T09:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.131594 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.131647 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.131668 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.131688 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.131702 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:11Z","lastTransitionTime":"2025-10-02T09:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.234508 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.234564 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.234595 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.234614 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.234624 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:11Z","lastTransitionTime":"2025-10-02T09:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.342938 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.342997 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.343007 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.343023 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.343036 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:11Z","lastTransitionTime":"2025-10-02T09:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.446214 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.446269 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.446279 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.446295 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.446304 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:11Z","lastTransitionTime":"2025-10-02T09:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.548796 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.548843 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.548856 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.548870 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.548879 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:11Z","lastTransitionTime":"2025-10-02T09:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.651633 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.651684 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.651699 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.651718 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.651730 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:11Z","lastTransitionTime":"2025-10-02T09:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.754697 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.754737 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.754748 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.754762 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.754771 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:11Z","lastTransitionTime":"2025-10-02T09:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.856967 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.857013 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.857025 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.857042 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.857053 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:11Z","lastTransitionTime":"2025-10-02T09:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.912329 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.912622 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.912660 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:50:11 crc kubenswrapper[4934]: E1002 09:50:11.912757 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.912826 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:50:11 crc kubenswrapper[4934]: E1002 09:50:11.912882 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:50:11 crc kubenswrapper[4934]: E1002 09:50:11.912955 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.913128 4934 scope.go:117] "RemoveContainer" containerID="358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe" Oct 02 09:50:11 crc kubenswrapper[4934]: E1002 09:50:11.913280 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:50:11 crc kubenswrapper[4934]: E1002 09:50:11.913332 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-th4v6_openshift-ovn-kubernetes(1d4841b5-0469-461e-875c-25b9fe848141)\"" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" podUID="1d4841b5-0469-461e-875c-25b9fe848141" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.960181 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.960225 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.960235 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.960252 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:11 crc kubenswrapper[4934]: I1002 09:50:11.960265 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:11Z","lastTransitionTime":"2025-10-02T09:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.038016 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:50:12 crc kubenswrapper[4934]: E1002 09:50:12.038279 4934 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:50:12 crc kubenswrapper[4934]: E1002 09:50:12.038416 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:51:16.038392607 +0000 UTC m=+147.791034159 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.062398 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.062440 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.062451 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.062471 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.062484 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:12Z","lastTransitionTime":"2025-10-02T09:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.138932 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.139060 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.139101 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.139147 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:50:12 crc kubenswrapper[4934]: E1002 09:50:12.139273 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:50:12 crc kubenswrapper[4934]: E1002 09:50:12.139290 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:50:12 crc kubenswrapper[4934]: E1002 09:50:12.139295 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:51:16.139259985 +0000 UTC m=+147.891901507 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:12 crc kubenswrapper[4934]: E1002 09:50:12.139301 4934 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:50:12 crc kubenswrapper[4934]: E1002 09:50:12.139344 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:50:12 crc kubenswrapper[4934]: E1002 09:50:12.139356 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:50:12 crc kubenswrapper[4934]: E1002 09:50:12.139365 4934 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:50:12 crc kubenswrapper[4934]: E1002 09:50:12.139383 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 09:51:16.139372907 +0000 UTC m=+147.892014439 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:50:12 crc kubenswrapper[4934]: E1002 09:50:12.139411 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 09:51:16.139393728 +0000 UTC m=+147.892035330 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:50:12 crc kubenswrapper[4934]: E1002 09:50:12.139448 4934 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:50:12 crc kubenswrapper[4934]: E1002 09:50:12.140274 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:51:16.13946325 +0000 UTC m=+147.892104862 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.164444 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.164479 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.164493 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.164509 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.164519 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:12Z","lastTransitionTime":"2025-10-02T09:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.271386 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.271434 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.271453 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.271481 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.271495 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:12Z","lastTransitionTime":"2025-10-02T09:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.374036 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.374086 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.374104 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.374126 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.374143 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:12Z","lastTransitionTime":"2025-10-02T09:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.477117 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.477191 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.477213 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.477240 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.477260 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:12Z","lastTransitionTime":"2025-10-02T09:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.580540 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.580690 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.580715 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.580745 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.580768 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:12Z","lastTransitionTime":"2025-10-02T09:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.683853 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.683894 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.683921 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.683947 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.683957 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:12Z","lastTransitionTime":"2025-10-02T09:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.787478 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.787529 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.787541 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.787556 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.787566 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:12Z","lastTransitionTime":"2025-10-02T09:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.890503 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.890550 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.890563 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.890827 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.890861 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:12Z","lastTransitionTime":"2025-10-02T09:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.925933 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.993381 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.993424 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.993441 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.993462 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:12 crc kubenswrapper[4934]: I1002 09:50:12.993478 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:12Z","lastTransitionTime":"2025-10-02T09:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.097886 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.097957 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.097972 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.098001 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.098018 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:13Z","lastTransitionTime":"2025-10-02T09:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.200815 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.200865 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.200875 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.200890 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.200899 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:13Z","lastTransitionTime":"2025-10-02T09:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.303442 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.303738 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.303842 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.303934 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.304018 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:13Z","lastTransitionTime":"2025-10-02T09:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.406688 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.406741 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.406758 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.406783 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.406802 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:13Z","lastTransitionTime":"2025-10-02T09:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.509001 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.509054 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.509063 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.509079 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.509091 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:13Z","lastTransitionTime":"2025-10-02T09:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.611806 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.611868 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.611883 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.611900 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.611911 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:13Z","lastTransitionTime":"2025-10-02T09:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.714119 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.714170 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.714182 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.714201 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.714212 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:13Z","lastTransitionTime":"2025-10-02T09:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.816444 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.816487 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.816499 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.816536 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.816551 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:13Z","lastTransitionTime":"2025-10-02T09:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.912911 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.913176 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.913195 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.913196 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:50:13 crc kubenswrapper[4934]: E1002 09:50:13.913428 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:50:13 crc kubenswrapper[4934]: E1002 09:50:13.913560 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:50:13 crc kubenswrapper[4934]: E1002 09:50:13.913689 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:50:13 crc kubenswrapper[4934]: E1002 09:50:13.913793 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.918826 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.918917 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.918979 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.919057 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:13 crc kubenswrapper[4934]: I1002 09:50:13.919121 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:13Z","lastTransitionTime":"2025-10-02T09:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.021847 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.022157 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.022441 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.022520 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.022607 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:14Z","lastTransitionTime":"2025-10-02T09:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.125981 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.126026 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.126039 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.126057 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.126071 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:14Z","lastTransitionTime":"2025-10-02T09:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.228878 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.228910 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.228919 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.228933 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.228943 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:14Z","lastTransitionTime":"2025-10-02T09:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.331543 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.331603 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.331612 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.331629 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.331638 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:14Z","lastTransitionTime":"2025-10-02T09:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.450100 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.450142 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.450153 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.450168 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.450177 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:14Z","lastTransitionTime":"2025-10-02T09:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.552729 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.553251 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.553364 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.553448 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.553521 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:14Z","lastTransitionTime":"2025-10-02T09:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.659487 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.659797 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.659908 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.660003 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.660079 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:14Z","lastTransitionTime":"2025-10-02T09:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.763340 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.763394 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.763409 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.763429 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.763446 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:14Z","lastTransitionTime":"2025-10-02T09:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.868018 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.868700 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.868802 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.868902 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.869017 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:14Z","lastTransitionTime":"2025-10-02T09:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.971527 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.971890 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.971977 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.972058 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:14 crc kubenswrapper[4934]: I1002 09:50:14.972315 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:14Z","lastTransitionTime":"2025-10-02T09:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.075250 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.075316 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.075338 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.075366 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.075385 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:15Z","lastTransitionTime":"2025-10-02T09:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.178173 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.178202 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.178211 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.178225 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.178234 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:15Z","lastTransitionTime":"2025-10-02T09:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.281474 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.281522 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.281540 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.281563 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.281624 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:15Z","lastTransitionTime":"2025-10-02T09:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.384246 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.384299 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.384314 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.384335 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.384351 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:15Z","lastTransitionTime":"2025-10-02T09:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.486988 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.487039 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.487052 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.487071 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.487083 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:15Z","lastTransitionTime":"2025-10-02T09:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.589507 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.589551 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.589563 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.589601 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.589617 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:15Z","lastTransitionTime":"2025-10-02T09:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.693308 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.693366 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.693385 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.693409 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.693429 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:15Z","lastTransitionTime":"2025-10-02T09:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.796557 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.796614 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.796627 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.796646 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.796662 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:15Z","lastTransitionTime":"2025-10-02T09:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.899112 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.899668 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.899750 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.899816 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.899877 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:15Z","lastTransitionTime":"2025-10-02T09:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.912694 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.912724 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.912759 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:50:15 crc kubenswrapper[4934]: I1002 09:50:15.912762 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:50:15 crc kubenswrapper[4934]: E1002 09:50:15.912929 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:50:15 crc kubenswrapper[4934]: E1002 09:50:15.913047 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:50:15 crc kubenswrapper[4934]: E1002 09:50:15.913208 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:50:15 crc kubenswrapper[4934]: E1002 09:50:15.913330 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.003154 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.003228 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.003247 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.003273 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.003290 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:16Z","lastTransitionTime":"2025-10-02T09:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.105766 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.105822 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.105832 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.105846 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.105856 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:16Z","lastTransitionTime":"2025-10-02T09:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.208795 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.208843 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.208859 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.208882 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.208899 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:16Z","lastTransitionTime":"2025-10-02T09:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.312247 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.312324 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.312351 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.312384 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.312407 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:16Z","lastTransitionTime":"2025-10-02T09:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.393978 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.394016 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.394027 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.394043 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.394054 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:16Z","lastTransitionTime":"2025-10-02T09:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:16 crc kubenswrapper[4934]: E1002 09:50:16.409697 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.414793 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.414843 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.414862 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.414886 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.414902 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:16Z","lastTransitionTime":"2025-10-02T09:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:16 crc kubenswrapper[4934]: E1002 09:50:16.433141 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.442281 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.442316 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.442327 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.442345 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.442356 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:16Z","lastTransitionTime":"2025-10-02T09:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:16 crc kubenswrapper[4934]: E1002 09:50:16.462157 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.469143 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.469445 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.469610 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.469775 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.470010 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:16Z","lastTransitionTime":"2025-10-02T09:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:16 crc kubenswrapper[4934]: E1002 09:50:16.485808 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.490626 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.490919 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.491050 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.491196 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.491335 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:16Z","lastTransitionTime":"2025-10-02T09:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:16 crc kubenswrapper[4934]: E1002 09:50:16.507569 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:16 crc kubenswrapper[4934]: E1002 09:50:16.507855 4934 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.510153 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.510254 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.510282 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.510311 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.510331 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:16Z","lastTransitionTime":"2025-10-02T09:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.612975 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.613018 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.613028 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.613043 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.613053 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:16Z","lastTransitionTime":"2025-10-02T09:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.715337 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.715417 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.715438 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.715457 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.715471 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:16Z","lastTransitionTime":"2025-10-02T09:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.818067 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.818127 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.818143 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.818169 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.818188 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:16Z","lastTransitionTime":"2025-10-02T09:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.920357 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.920403 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.920418 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.920438 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:16 crc kubenswrapper[4934]: I1002 09:50:16.920453 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:16Z","lastTransitionTime":"2025-10-02T09:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.022656 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.022715 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.022728 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.022748 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.022761 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:17Z","lastTransitionTime":"2025-10-02T09:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.125711 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.125760 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.125772 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.125789 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.125801 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:17Z","lastTransitionTime":"2025-10-02T09:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.228759 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.228822 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.228831 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.228847 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.228856 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:17Z","lastTransitionTime":"2025-10-02T09:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.331769 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.331820 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.331833 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.331849 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.331859 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:17Z","lastTransitionTime":"2025-10-02T09:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.434220 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.434272 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.434284 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.434307 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.434342 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:17Z","lastTransitionTime":"2025-10-02T09:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.537217 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.537259 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.537269 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.537287 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.537296 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:17Z","lastTransitionTime":"2025-10-02T09:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.640293 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.640400 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.640419 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.640446 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.640463 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:17Z","lastTransitionTime":"2025-10-02T09:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.743319 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.743385 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.743402 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.743430 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.743450 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:17Z","lastTransitionTime":"2025-10-02T09:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.846251 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.846298 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.846311 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.846331 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.846343 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:17Z","lastTransitionTime":"2025-10-02T09:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.912794 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.912813 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.912841 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.912863 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:50:17 crc kubenswrapper[4934]: E1002 09:50:17.913953 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:50:17 crc kubenswrapper[4934]: E1002 09:50:17.914007 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:50:17 crc kubenswrapper[4934]: E1002 09:50:17.914072 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:50:17 crc kubenswrapper[4934]: E1002 09:50:17.914136 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.949370 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.949440 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.949458 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.949483 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:17 crc kubenswrapper[4934]: I1002 09:50:17.949507 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:17Z","lastTransitionTime":"2025-10-02T09:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.051945 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.052016 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.052036 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.052061 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.052115 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:18Z","lastTransitionTime":"2025-10-02T09:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.154879 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.155560 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.155635 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.155659 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.155669 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:18Z","lastTransitionTime":"2025-10-02T09:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.259035 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.259063 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.259072 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.259084 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.259092 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:18Z","lastTransitionTime":"2025-10-02T09:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.361433 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.361491 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.361506 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.361523 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.361536 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:18Z","lastTransitionTime":"2025-10-02T09:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.464442 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.464509 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.464519 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.464534 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.464546 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:18Z","lastTransitionTime":"2025-10-02T09:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.570365 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.570404 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.570414 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.570436 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.570447 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:18Z","lastTransitionTime":"2025-10-02T09:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.673395 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.673452 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.673474 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.673504 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.673527 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:18Z","lastTransitionTime":"2025-10-02T09:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.776647 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.776840 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.776860 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.776901 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.776917 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:18Z","lastTransitionTime":"2025-10-02T09:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.879978 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.880017 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.880026 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.880065 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.880080 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:18Z","lastTransitionTime":"2025-10-02T09:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.933310 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.946107 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.958895 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.969021 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7186037-ba99-4a2b-8a22-6417ffc44b80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407acc937940ce03f746be72752c5efa2ce7e416c39a80c2763befdec18daa56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f14dcea7cf5a10dfe10430bfd3d6e41710d33685f43cdc634206a17bbb3697f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9af6ae65b40a9786691a168f58a28aa4e1f1fdb401976046e4657bbbc0d44fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6711e8eee36c5d872a158ccedc38e3739b94e96e19564babf96bd6fef2dad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c6711e8eee36c5d872a158ccedc38e3739b94e96e19564babf96bd6fef2dad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.979133 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.984719 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.984775 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.984792 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.984814 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.984832 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:18Z","lastTransitionTime":"2025-10-02T09:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:18 crc kubenswrapper[4934]: I1002 09:50:18.991444 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:18Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.017189 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"message\\\":\\\" EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:9154, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}, services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_UDP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"UDP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}}\\\\nI1002 09:49:55.844402 6884 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 09:49:55.844456 6884 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-th4v6_openshift-ovn-kubernetes(1d4841b5-0469-461e-875c-25b9fe848141)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.030035 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://106b1f23a09ef32bc00222f00777a5a8c211b61c6f3d9fc4b67d7708f8091bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fc3a7a93196f00272230d3c06af1b408d42a806a5e65fbb3cf131ddfb37077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ffb9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.045377 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27680347ccabc1c590a7ffc59feeb56ad6123a3ae83806a893ac9283d4c6b5ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.059339 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.072451 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.100965 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.108981 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.109040 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.109057 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.109081 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.109099 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:19Z","lastTransitionTime":"2025-10-02T09:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.117712 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9126a401-adbd-4a42-a437-b76fa3fbcb10\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab08caf3ec6f4cec18fc54615059138c4110223b1413c5fbbc1cf718dade62f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0b4710144b27bd0899de27f526502d6b23a51bcffa2ecd7a41efd245fff5da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0b4710144b27bd0899de27f526502d6b23a51bcffa2ecd7a41efd245fff5da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.143458 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.161549 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851e2f96cda63332af97a90a26088ecdd7fcf2d0498b53e5d542aa532b048e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.177064 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.192468 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.204282 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84da7560cede71d06bd39eeee26703d00479335105cce31a29648fbd36512917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:57Z\\\",\\\"message\\\":\\\"2025-10-02T09:49:12+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_730f868f-8db1-47c0-9b6f-91623ab15850\\\\n2025-10-02T09:49:12+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_730f868f-8db1-47c0-9b6f-91623ab15850 to /host/opt/cni/bin/\\\\n2025-10-02T09:49:12Z [verbose] multus-daemon started\\\\n2025-10-02T09:49:12Z [verbose] Readiness Indicator file check\\\\n2025-10-02T09:49:57Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.211333 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.211373 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.211381 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.211397 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.211407 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:19Z","lastTransitionTime":"2025-10-02T09:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.215351 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7cgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7cgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:19Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.314143 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.314230 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.314241 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.314257 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.314267 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:19Z","lastTransitionTime":"2025-10-02T09:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.415777 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.415853 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.415869 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.415891 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.415948 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:19Z","lastTransitionTime":"2025-10-02T09:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.518746 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.518798 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.518814 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.518834 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.518850 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:19Z","lastTransitionTime":"2025-10-02T09:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.621688 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.621742 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.621755 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.621774 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.621786 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:19Z","lastTransitionTime":"2025-10-02T09:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.724099 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.724153 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.724165 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.724182 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.724195 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:19Z","lastTransitionTime":"2025-10-02T09:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.827313 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.827358 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.827369 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.827384 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.827395 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:19Z","lastTransitionTime":"2025-10-02T09:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.912862 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.912923 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:50:19 crc kubenswrapper[4934]: E1002 09:50:19.913000 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.912944 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:50:19 crc kubenswrapper[4934]: E1002 09:50:19.913174 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:50:19 crc kubenswrapper[4934]: E1002 09:50:19.913401 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.913662 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:50:19 crc kubenswrapper[4934]: E1002 09:50:19.913802 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.931144 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.931213 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.931237 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.931267 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:19 crc kubenswrapper[4934]: I1002 09:50:19.931288 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:19Z","lastTransitionTime":"2025-10-02T09:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.035763 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.035834 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.035856 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.035884 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.035903 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:20Z","lastTransitionTime":"2025-10-02T09:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.138059 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.138130 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.138147 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.138625 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.138683 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:20Z","lastTransitionTime":"2025-10-02T09:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.241069 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.241103 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.241111 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.241123 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.241131 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:20Z","lastTransitionTime":"2025-10-02T09:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.343270 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.343332 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.343341 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.343357 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.343366 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:20Z","lastTransitionTime":"2025-10-02T09:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.446161 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.446218 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.446233 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.446250 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.446262 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:20Z","lastTransitionTime":"2025-10-02T09:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.549221 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.549257 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.549266 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.549280 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.549289 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:20Z","lastTransitionTime":"2025-10-02T09:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.651374 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.651423 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.651435 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.651471 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.651486 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:20Z","lastTransitionTime":"2025-10-02T09:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.753658 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.753701 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.753711 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.753724 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.753734 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:20Z","lastTransitionTime":"2025-10-02T09:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.856602 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.856659 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.856674 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.856694 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.856710 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:20Z","lastTransitionTime":"2025-10-02T09:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.958961 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.959002 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.959013 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.959027 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:20 crc kubenswrapper[4934]: I1002 09:50:20.959038 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:20Z","lastTransitionTime":"2025-10-02T09:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.061852 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.061891 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.061900 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.061914 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.061924 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:21Z","lastTransitionTime":"2025-10-02T09:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.165485 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.165530 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.165550 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.165571 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.165614 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:21Z","lastTransitionTime":"2025-10-02T09:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.268748 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.268786 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.268795 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.268826 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.268834 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:21Z","lastTransitionTime":"2025-10-02T09:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.371857 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.371907 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.371917 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.371933 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.371944 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:21Z","lastTransitionTime":"2025-10-02T09:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.475169 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.475254 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.475272 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.475295 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.475312 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:21Z","lastTransitionTime":"2025-10-02T09:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.577997 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.578062 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.578079 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.578104 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.578121 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:21Z","lastTransitionTime":"2025-10-02T09:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.681109 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.681175 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.681199 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.681227 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.681250 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:21Z","lastTransitionTime":"2025-10-02T09:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.784093 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.784145 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.784157 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.784176 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.784189 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:21Z","lastTransitionTime":"2025-10-02T09:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.887278 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.887327 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.887339 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.887359 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.887373 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:21Z","lastTransitionTime":"2025-10-02T09:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.913112 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.913116 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:50:21 crc kubenswrapper[4934]: E1002 09:50:21.913366 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.913162 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.913138 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:50:21 crc kubenswrapper[4934]: E1002 09:50:21.913486 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:50:21 crc kubenswrapper[4934]: E1002 09:50:21.913840 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:50:21 crc kubenswrapper[4934]: E1002 09:50:21.913984 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.990043 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.990101 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.990141 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.990170 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:21 crc kubenswrapper[4934]: I1002 09:50:21.990190 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:21Z","lastTransitionTime":"2025-10-02T09:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.093519 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.093560 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.093595 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.093616 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.093628 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:22Z","lastTransitionTime":"2025-10-02T09:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.196081 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.196151 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.196168 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.196190 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.196207 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:22Z","lastTransitionTime":"2025-10-02T09:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.299262 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.299302 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.299311 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.299326 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.299336 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:22Z","lastTransitionTime":"2025-10-02T09:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.402264 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.402309 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.402319 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.402334 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.402344 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:22Z","lastTransitionTime":"2025-10-02T09:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.504326 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.504358 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.504366 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.504379 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.504387 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:22Z","lastTransitionTime":"2025-10-02T09:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.606764 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.606839 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.606851 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.606866 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.606878 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:22Z","lastTransitionTime":"2025-10-02T09:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.709996 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.710043 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.710054 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.710071 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.710082 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:22Z","lastTransitionTime":"2025-10-02T09:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.812491 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.812532 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.812543 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.812560 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.812572 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:22Z","lastTransitionTime":"2025-10-02T09:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.915124 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.915196 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.915208 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.915228 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:22 crc kubenswrapper[4934]: I1002 09:50:22.915240 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:22Z","lastTransitionTime":"2025-10-02T09:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.017286 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.017330 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.017341 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.017357 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.017368 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:23Z","lastTransitionTime":"2025-10-02T09:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.120069 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.120123 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.120141 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.120165 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.120182 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:23Z","lastTransitionTime":"2025-10-02T09:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.222853 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.222913 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.222926 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.222947 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.222963 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:23Z","lastTransitionTime":"2025-10-02T09:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.325321 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.325356 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.325368 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.325385 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.325394 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:23Z","lastTransitionTime":"2025-10-02T09:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.427869 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.427944 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.427956 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.427973 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.427985 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:23Z","lastTransitionTime":"2025-10-02T09:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.530946 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.530996 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.531006 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.531021 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.531030 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:23Z","lastTransitionTime":"2025-10-02T09:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.633255 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.633299 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.633312 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.633331 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.633343 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:23Z","lastTransitionTime":"2025-10-02T09:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.735907 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.735945 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.735953 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.735968 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.735978 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:23Z","lastTransitionTime":"2025-10-02T09:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.838203 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.838248 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.838294 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.838311 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.838325 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:23Z","lastTransitionTime":"2025-10-02T09:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.912335 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.912423 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:50:23 crc kubenswrapper[4934]: E1002 09:50:23.912464 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.912647 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:50:23 crc kubenswrapper[4934]: E1002 09:50:23.912657 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:50:23 crc kubenswrapper[4934]: E1002 09:50:23.912690 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.912733 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:50:23 crc kubenswrapper[4934]: E1002 09:50:23.912773 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.941705 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.941755 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.941770 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.941790 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:23 crc kubenswrapper[4934]: I1002 09:50:23.941803 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:23Z","lastTransitionTime":"2025-10-02T09:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.044162 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.044221 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.044234 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.044253 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.044269 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:24Z","lastTransitionTime":"2025-10-02T09:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.146930 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.146985 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.146995 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.147010 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.147020 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:24Z","lastTransitionTime":"2025-10-02T09:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.250055 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.250107 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.250119 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.250137 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.250147 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:24Z","lastTransitionTime":"2025-10-02T09:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.352668 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.352702 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.352712 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.352727 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.352739 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:24Z","lastTransitionTime":"2025-10-02T09:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.455694 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.455744 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.455755 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.455772 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.455785 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:24Z","lastTransitionTime":"2025-10-02T09:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.558106 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.558149 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.558159 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.558179 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.558191 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:24Z","lastTransitionTime":"2025-10-02T09:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.660456 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.660499 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.660509 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.660526 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.660539 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:24Z","lastTransitionTime":"2025-10-02T09:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.763342 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.763396 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.763413 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.763434 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.763450 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:24Z","lastTransitionTime":"2025-10-02T09:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.867405 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.867455 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.867466 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.867482 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.867496 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:24Z","lastTransitionTime":"2025-10-02T09:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.969498 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.969533 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.969540 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.969554 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:24 crc kubenswrapper[4934]: I1002 09:50:24.969563 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:24Z","lastTransitionTime":"2025-10-02T09:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.071427 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.071465 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.071476 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.071492 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.071505 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:25Z","lastTransitionTime":"2025-10-02T09:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.173954 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.173995 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.174005 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.174021 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.174033 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:25Z","lastTransitionTime":"2025-10-02T09:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.277721 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.277935 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.277960 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.277984 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.278001 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:25Z","lastTransitionTime":"2025-10-02T09:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.380210 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.380265 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.380282 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.380305 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.380320 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:25Z","lastTransitionTime":"2025-10-02T09:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.482144 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.482188 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.482198 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.482212 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.482227 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:25Z","lastTransitionTime":"2025-10-02T09:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.584628 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.584674 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.584685 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.584701 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.584712 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:25Z","lastTransitionTime":"2025-10-02T09:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.687916 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.687978 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.687990 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.688007 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.688018 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:25Z","lastTransitionTime":"2025-10-02T09:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.789918 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.789964 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.789983 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.790008 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.790024 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:25Z","lastTransitionTime":"2025-10-02T09:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.892938 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.892971 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.892981 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.892997 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.893009 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:25Z","lastTransitionTime":"2025-10-02T09:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.912132 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.912173 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.912192 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:50:25 crc kubenswrapper[4934]: E1002 09:50:25.912262 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.912145 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:50:25 crc kubenswrapper[4934]: E1002 09:50:25.912724 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:50:25 crc kubenswrapper[4934]: E1002 09:50:25.912987 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:50:25 crc kubenswrapper[4934]: E1002 09:50:25.913092 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.995295 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.995642 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.995760 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.995847 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:25 crc kubenswrapper[4934]: I1002 09:50:25.995932 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:25Z","lastTransitionTime":"2025-10-02T09:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.098688 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.098946 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.099006 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.099077 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.099134 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:26Z","lastTransitionTime":"2025-10-02T09:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.201562 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.201659 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.201679 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.201706 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.201724 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:26Z","lastTransitionTime":"2025-10-02T09:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.305236 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.305492 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.305627 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.305723 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.305785 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:26Z","lastTransitionTime":"2025-10-02T09:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.408772 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.408814 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.408826 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.408844 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.408856 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:26Z","lastTransitionTime":"2025-10-02T09:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.511815 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.511866 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.511877 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.511893 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.511905 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:26Z","lastTransitionTime":"2025-10-02T09:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.615827 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.615895 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.615919 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.615947 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.615971 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:26Z","lastTransitionTime":"2025-10-02T09:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.719291 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.719364 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.719388 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.719418 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.719437 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:26Z","lastTransitionTime":"2025-10-02T09:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.760678 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.760738 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.760756 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.760783 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.760801 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:26Z","lastTransitionTime":"2025-10-02T09:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:26 crc kubenswrapper[4934]: E1002 09:50:26.778823 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.783820 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.783865 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.783883 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.783907 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.783925 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:26Z","lastTransitionTime":"2025-10-02T09:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:26 crc kubenswrapper[4934]: E1002 09:50:26.807538 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.812414 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.812463 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.812480 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.812503 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.812521 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:26Z","lastTransitionTime":"2025-10-02T09:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:26 crc kubenswrapper[4934]: E1002 09:50:26.830557 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.834789 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.834839 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.834854 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.834872 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.834884 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:26Z","lastTransitionTime":"2025-10-02T09:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:26 crc kubenswrapper[4934]: E1002 09:50:26.849060 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.854011 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.854046 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.854059 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.854081 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.854095 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:26Z","lastTransitionTime":"2025-10-02T09:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:26 crc kubenswrapper[4934]: E1002 09:50:26.868773 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:50:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cbe6531b-0669-4b46-873b-a08a7b00ffa5\\\",\\\"systemUUID\\\":\\\"a1c4609c-a340-4b44-bdf6-dcf5b8eeda7b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:26 crc kubenswrapper[4934]: E1002 09:50:26.868960 4934 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.870872 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.870901 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.870912 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.870930 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.870943 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:26Z","lastTransitionTime":"2025-10-02T09:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.914132 4934 scope.go:117] "RemoveContainer" containerID="358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe" Oct 02 09:50:26 crc kubenswrapper[4934]: E1002 09:50:26.914327 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-th4v6_openshift-ovn-kubernetes(1d4841b5-0469-461e-875c-25b9fe848141)\"" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" podUID="1d4841b5-0469-461e-875c-25b9fe848141" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.973945 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.973992 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.974003 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.974018 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:26 crc kubenswrapper[4934]: I1002 09:50:26.974027 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:26Z","lastTransitionTime":"2025-10-02T09:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.077075 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.077123 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.077134 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.077152 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.077163 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:27Z","lastTransitionTime":"2025-10-02T09:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.181264 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.181341 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.181354 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.181377 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.181404 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:27Z","lastTransitionTime":"2025-10-02T09:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.284527 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.284618 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.284645 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.284675 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.284693 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:27Z","lastTransitionTime":"2025-10-02T09:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.388179 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.388234 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.388244 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.388267 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.388282 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:27Z","lastTransitionTime":"2025-10-02T09:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.491963 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.492043 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.492059 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.492086 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.492104 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:27Z","lastTransitionTime":"2025-10-02T09:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.595170 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.595242 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.595264 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.595296 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.595317 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:27Z","lastTransitionTime":"2025-10-02T09:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.698335 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.698390 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.698405 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.698425 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.698441 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:27Z","lastTransitionTime":"2025-10-02T09:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.801459 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.801505 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.801518 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.801537 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.801551 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:27Z","lastTransitionTime":"2025-10-02T09:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.904410 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.904465 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.904486 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.904513 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.904531 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:27Z","lastTransitionTime":"2025-10-02T09:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.907381 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-metrics-certs\") pod \"network-metrics-daemon-n7cgz\" (UID: \"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\") " pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:50:27 crc kubenswrapper[4934]: E1002 09:50:27.907546 4934 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:50:27 crc kubenswrapper[4934]: E1002 09:50:27.907636 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-metrics-certs podName:0cbef3ee-c66f-47fa-94d3-5ce9892c403f nodeName:}" failed. No retries permitted until 2025-10-02 09:51:31.907617461 +0000 UTC m=+163.660258983 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-metrics-certs") pod "network-metrics-daemon-n7cgz" (UID: "0cbef3ee-c66f-47fa-94d3-5ce9892c403f") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.912302 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.912349 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.912358 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:50:27 crc kubenswrapper[4934]: I1002 09:50:27.912610 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:50:27 crc kubenswrapper[4934]: E1002 09:50:27.912604 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:50:27 crc kubenswrapper[4934]: E1002 09:50:27.912734 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:50:27 crc kubenswrapper[4934]: E1002 09:50:27.912858 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:50:27 crc kubenswrapper[4934]: E1002 09:50:27.913033 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.008269 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.008336 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.008359 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.008388 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.008406 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:28Z","lastTransitionTime":"2025-10-02T09:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.111216 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.111296 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.111306 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.111325 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.111335 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:28Z","lastTransitionTime":"2025-10-02T09:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.214821 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.214886 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.214910 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.214987 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.215011 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:28Z","lastTransitionTime":"2025-10-02T09:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.317727 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.317771 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.317782 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.317800 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.317812 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:28Z","lastTransitionTime":"2025-10-02T09:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.421265 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.421360 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.421379 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.421405 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.421420 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:28Z","lastTransitionTime":"2025-10-02T09:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.524739 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.524800 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.524810 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.524830 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.524843 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:28Z","lastTransitionTime":"2025-10-02T09:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.627650 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.627705 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.627728 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.627749 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.627763 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:28Z","lastTransitionTime":"2025-10-02T09:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.730759 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.730807 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.730822 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.730843 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.730855 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:28Z","lastTransitionTime":"2025-10-02T09:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.832997 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.833032 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.833042 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.833055 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.833064 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:28Z","lastTransitionTime":"2025-10-02T09:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.932847 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9126a401-adbd-4a42-a437-b76fa3fbcb10\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab08caf3ec6f4cec18fc54615059138c4110223b1413c5fbbc1cf718dade62f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0b4710144b27bd0899de27f526502d6b23a51bcffa2ecd7a41efd245fff5da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f0b4710144b27bd0899de27f526502d6b23a51bcffa2ecd7a41efd245fff5da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:28Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.939049 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.939119 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.939143 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.939173 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.939195 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:28Z","lastTransitionTime":"2025-10-02T09:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.963035 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d80b3f30-b3e5-4f8b-b75d-800b2e0063c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6f6e6d31d957728bd528f73ddaa79aac63d1b5eae4394ff8dc3a38014fdde0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d83711d1a7b613f313470070e404c610bd7de065be43a111c04328b3d686b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a33efb57c57f2c4493d65726e3213fb8d1512ecbbf2f30add1766de7045a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8727ef19a7c6cbfc9295d57a94a39b881630ecbaa068314c056b8fa0fe60c51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a118569d19754d23497808ab25d6fff5300735af5af91217505e2cb2132f566\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ec6990c2aae67c2a31da95dc0044881f41abac320f8ca3a16c7f323ced681ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://96eb11576e0984ef532ead362d94a8ffb7d1b874c0722811f100408cd7057eec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d01abdc91d1845913d2d892e521465522161b6e80280c71fabf0cb705c476b34\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:28Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.981416 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ds45z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73cb6891-fe6c-466e-a8b3-5497f28d2741\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://851e2f96cda63332af97a90a26088ecdd7fcf2d0498b53e5d542aa532b048e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880226a1c367cc7431ece6ff8140eecbd762ac60ba0fed1ba17e82bc7cc67ae8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0716d177fe3473b795af55c9dad5e99546492b142671143b3038e0734e89e58a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48dc5447f9adb1eb2b57ff2b441c32331d42b8075c8cebbfd9e8037592624ec7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efdc9193a8f71fdb003e6d4f3bbbb554645ccac6a528b1d1e2948ff6d0718495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea141896aad01c261174bb8e744db9d96c0167ed97b4843d58db354fa53a897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ad326dad3e3a64f9d93b259cf62d39f280cc6c175df85803cf15caca0da8e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxs8x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ds45z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:28Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:28 crc kubenswrapper[4934]: I1002 09:50:28.999756 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a88a3f1dde1d083854e34d20d2394cfa1e42c0003c02d798c7e2510176c77df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:28Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.017314 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.036894 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-79fxg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73968247-67dd-48cc-88a1-64afac657412\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84da7560cede71d06bd39eeee26703d00479335105cce31a29648fbd36512917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:57Z\\\",\\\"message\\\":\\\"2025-10-02T09:49:12+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_730f868f-8db1-47c0-9b6f-91623ab15850\\\\n2025-10-02T09:49:12+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_730f868f-8db1-47c0-9b6f-91623ab15850 to /host/opt/cni/bin/\\\\n2025-10-02T09:49:12Z [verbose] multus-daemon started\\\\n2025-10-02T09:49:12Z [verbose] Readiness Indicator file check\\\\n2025-10-02T09:49:57Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kmk48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-79fxg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.041466 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.041519 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.041531 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.041549 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.041561 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:29Z","lastTransitionTime":"2025-10-02T09:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.052394 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7cgz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvdrf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7cgz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.068724 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ee8dd367d4061bf3f0df18a5a88d5284b6c81dc9fbc08fe75bdca3adc78709d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.082711 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tctph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13155c94-2cc0-483d-afa0-68f2415404b0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec75952da7f3ed34697051a25efd19c466605567793cb25e841fa23066aed52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd7nm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:12Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tctph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.096722 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3100ca9b-475e-4b84-a266-42b2e228fe34\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff7c515e962c937c06f2d0e401947d9d12b38c5f6a4d5db3d81fab05f4cd13f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://376c355964456b573d8567bb4f8314b9f8ff0537c6dc7267ee72d4b7b7d99182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e1b93bb134d1b904e5453b25eb925e11e61045fb80b2ee829fdfef8e5e9e9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://234e1186712186af327a491c3f2af7e4cabd02b42d3c988c96c68e6f321dc2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.109743 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d7186037-ba99-4a2b-8a22-6417ffc44b80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407acc937940ce03f746be72752c5efa2ce7e416c39a80c2763befdec18daa56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f14dcea7cf5a10dfe10430bfd3d6e41710d33685f43cdc634206a17bbb3697f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9af6ae65b40a9786691a168f58a28aa4e1f1fdb401976046e4657bbbc0d44fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6711e8eee36c5d872a158ccedc38e3739b94e96e19564babf96bd6fef2dad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c6711e8eee36c5d872a158ccedc38e3739b94e96e19564babf96bd6fef2dad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.126984 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.142835 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62396656c07859ac7dcbddde08f0916475c79ba466955a96aaaa6f4dc35a366e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6c1aacebb004a9e04aa33d60b32487a7ce8beb18077d37f33738763330a056c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.144084 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.144225 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.144383 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.144530 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.144691 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:29Z","lastTransitionTime":"2025-10-02T09:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.168563 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d4841b5-0469-461e-875c-25b9fe848141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:49:55Z\\\",\\\"message\\\":\\\" EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:9154, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}, services.LB{Name:\\\\\\\"Service_openshift-dns/dns-default_UDP_node_router+switch_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"UDP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-dns/dns-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.10\\\\\\\", Port:53, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string(nil)}}\\\\nI1002 09:49:55.844402 6884 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 09:49:55.844456 6884 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-th4v6_openshift-ovn-kubernetes(1d4841b5-0469-461e-875c-25b9fe848141)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4mgrl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-th4v6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.185382 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c9999d1-d367-45d6-8c41-c3bc3ab23df4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://106b1f23a09ef32bc00222f00777a5a8c211b61c6f3d9fc4b67d7708f8091bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87fc3a7a93196f00272230d3c06af1b408d42a806a5e65fbb3cf131ddfb37077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7bpk6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ffb9r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.207129 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cf0a06e-1228-403e-ab26-a79621ba30ed\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:48:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc1ae11547e45869bc8f467dcbfe024a320071b60df0adeea0d5bcf886ee91ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1a829519d1e2aa8506bcda7cd98fba6d61c5867b98c48c2a7ee4968f247147c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c3464f5432936449b79b730a0c9fba6e841d0ddf1152055e3bf25879d683cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27680347ccabc1c590a7ffc59feeb56ad6123a3ae83806a893ac9283d4c6b5ac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a5d210d0bcaada693882184f5c00a8eb014a6175081a0fc02d73fad6072f97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:49:08Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 09:49:08.478918 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:49:08.479088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:49:08.479870 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-466428714/tls.crt::/tmp/serving-cert-466428714/tls.key\\\\\\\"\\\\nI1002 09:49:08.958410 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:49:08.966351 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:49:08.966377 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:49:08.966396 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:49:08.966401 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:49:08.973036 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1002 09:49:08.973057 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 09:49:08.973067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973073 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:49:08.973080 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:49:08.973084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:49:08.973088 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:49:08.973091 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:49:08.974823 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:49:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3ce74bec72af8b4b48dc2ecd0fb2d74376fb1adacb1bd6f6799bcc7255797a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:48:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a90847d883f16e6c4e53cc3c63debf9d082a25ea90f18c669e0a3adadc49b47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:48:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.227515 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:07Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.239028 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-85mph" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e2b3511-723e-4cb7-aaf6-1b077ae88458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51ec9bcc68ef7239ed68398925fb5d52d8dfca2a9e5d00fd06b42a0dae8ac36e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lf88n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-85mph\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.302158 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.302189 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.302197 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.302210 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.302218 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:29Z","lastTransitionTime":"2025-10-02T09:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.304199 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71db06ef-05b0-4f58-b251-b27117a8500a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:49:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d01d6ecc0ccdb47054c51dfb39fa78ccdd102dd2dadf6b58d66f8081de9a119a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-stfqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:49:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-djh5z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:50:29Z is after 2025-08-24T17:21:41Z" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.404537 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.404607 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.404619 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.404636 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.404649 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:29Z","lastTransitionTime":"2025-10-02T09:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.507081 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.507139 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.507155 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.507175 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.507190 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:29Z","lastTransitionTime":"2025-10-02T09:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.615721 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.615830 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.615849 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.615878 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.615891 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:29Z","lastTransitionTime":"2025-10-02T09:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.719767 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.719822 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.719833 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.719852 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.719865 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:29Z","lastTransitionTime":"2025-10-02T09:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.822662 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.822716 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.822729 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.822748 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.822760 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:29Z","lastTransitionTime":"2025-10-02T09:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.912311 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.912381 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.912386 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:50:29 crc kubenswrapper[4934]: E1002 09:50:29.912472 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.912497 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:50:29 crc kubenswrapper[4934]: E1002 09:50:29.912666 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:50:29 crc kubenswrapper[4934]: E1002 09:50:29.912820 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:50:29 crc kubenswrapper[4934]: E1002 09:50:29.912886 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.925700 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.925742 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.925754 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.925772 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:29 crc kubenswrapper[4934]: I1002 09:50:29.925784 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:29Z","lastTransitionTime":"2025-10-02T09:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.029106 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.029174 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.029187 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.029212 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.029225 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:30Z","lastTransitionTime":"2025-10-02T09:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.132914 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.132949 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.132957 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.132971 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.132980 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:30Z","lastTransitionTime":"2025-10-02T09:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.235308 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.235345 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.235354 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.235368 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.235380 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:30Z","lastTransitionTime":"2025-10-02T09:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.338774 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.338841 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.338860 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.338888 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.338907 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:30Z","lastTransitionTime":"2025-10-02T09:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.441157 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.441214 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.441230 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.441250 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.441266 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:30Z","lastTransitionTime":"2025-10-02T09:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.544180 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.544306 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.544332 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.544366 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.544390 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:30Z","lastTransitionTime":"2025-10-02T09:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.647882 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.647955 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.648021 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.648055 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.648079 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:30Z","lastTransitionTime":"2025-10-02T09:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.750783 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.750854 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.750869 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.750886 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.750897 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:30Z","lastTransitionTime":"2025-10-02T09:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.853851 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.853929 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.853947 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.853972 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.853992 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:30Z","lastTransitionTime":"2025-10-02T09:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.957835 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.957872 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.957880 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.957893 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:30 crc kubenswrapper[4934]: I1002 09:50:30.957902 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:30Z","lastTransitionTime":"2025-10-02T09:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.059930 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.059975 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.059988 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.060006 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.060018 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:31Z","lastTransitionTime":"2025-10-02T09:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.163144 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.163193 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.163208 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.163231 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.163246 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:31Z","lastTransitionTime":"2025-10-02T09:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.266342 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.266398 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.266412 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.266438 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.266453 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:31Z","lastTransitionTime":"2025-10-02T09:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.369697 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.369776 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.369795 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.369818 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.369837 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:31Z","lastTransitionTime":"2025-10-02T09:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.472923 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.472981 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.473000 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.473028 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.473047 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:31Z","lastTransitionTime":"2025-10-02T09:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.576794 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.577107 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.577359 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.577625 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.577885 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:31Z","lastTransitionTime":"2025-10-02T09:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.680973 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.681025 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.681042 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.681065 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.681085 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:31Z","lastTransitionTime":"2025-10-02T09:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.784763 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.785419 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.785841 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.786034 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.786162 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:31Z","lastTransitionTime":"2025-10-02T09:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.889843 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.890284 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.890357 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.890481 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.890563 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:31Z","lastTransitionTime":"2025-10-02T09:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.913089 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.913193 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.913213 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:50:31 crc kubenswrapper[4934]: E1002 09:50:31.913327 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.913389 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:50:31 crc kubenswrapper[4934]: E1002 09:50:31.913527 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:50:31 crc kubenswrapper[4934]: E1002 09:50:31.913762 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:50:31 crc kubenswrapper[4934]: E1002 09:50:31.913850 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.993916 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.993970 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.993982 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.994001 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:31 crc kubenswrapper[4934]: I1002 09:50:31.994013 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:31Z","lastTransitionTime":"2025-10-02T09:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.096780 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.096850 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.096884 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.096913 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.096934 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:32Z","lastTransitionTime":"2025-10-02T09:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.198833 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.199035 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.199048 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.199063 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.199073 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:32Z","lastTransitionTime":"2025-10-02T09:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.301934 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.301995 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.302007 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.302027 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.302040 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:32Z","lastTransitionTime":"2025-10-02T09:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.404630 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.404780 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.404811 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.404849 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.404868 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:32Z","lastTransitionTime":"2025-10-02T09:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.507255 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.507309 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.507322 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.507339 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.507351 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:32Z","lastTransitionTime":"2025-10-02T09:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.609881 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.609932 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.609940 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.609956 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.609966 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:32Z","lastTransitionTime":"2025-10-02T09:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.712788 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.712846 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.712858 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.712877 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.712891 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:32Z","lastTransitionTime":"2025-10-02T09:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.815896 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.815952 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.815963 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.815986 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.816000 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:32Z","lastTransitionTime":"2025-10-02T09:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.920997 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.921077 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.921115 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.921779 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:32 crc kubenswrapper[4934]: I1002 09:50:32.921865 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:32Z","lastTransitionTime":"2025-10-02T09:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.026434 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.026495 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.026512 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.026532 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.026546 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:33Z","lastTransitionTime":"2025-10-02T09:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.129370 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.129427 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.129444 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.129466 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.129480 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:33Z","lastTransitionTime":"2025-10-02T09:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.233274 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.233361 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.233384 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.233414 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.233435 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:33Z","lastTransitionTime":"2025-10-02T09:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.336186 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.336240 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.336253 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.336276 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.336287 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:33Z","lastTransitionTime":"2025-10-02T09:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.440538 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.440622 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.440637 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.440659 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.440673 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:33Z","lastTransitionTime":"2025-10-02T09:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.544053 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.544111 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.544128 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.544153 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.544176 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:33Z","lastTransitionTime":"2025-10-02T09:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.647748 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.647820 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.647847 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.647885 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.647907 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:33Z","lastTransitionTime":"2025-10-02T09:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.750960 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.751004 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.751019 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.751042 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.751058 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:33Z","lastTransitionTime":"2025-10-02T09:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.853112 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.853343 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.853429 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.853493 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.853555 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:33Z","lastTransitionTime":"2025-10-02T09:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.913191 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:50:33 crc kubenswrapper[4934]: E1002 09:50:33.913380 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.913479 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.913569 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:50:33 crc kubenswrapper[4934]: E1002 09:50:33.913716 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.913726 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:50:33 crc kubenswrapper[4934]: E1002 09:50:33.913947 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:50:33 crc kubenswrapper[4934]: E1002 09:50:33.914221 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.957077 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.957117 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.957128 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.957143 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:33 crc kubenswrapper[4934]: I1002 09:50:33.957155 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:33Z","lastTransitionTime":"2025-10-02T09:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.059378 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.059429 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.059442 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.059457 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.059467 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:34Z","lastTransitionTime":"2025-10-02T09:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.162361 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.162403 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.162411 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.162425 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.162435 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:34Z","lastTransitionTime":"2025-10-02T09:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.265068 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.265150 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.265164 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.265187 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.265201 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:34Z","lastTransitionTime":"2025-10-02T09:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.368383 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.368427 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.368438 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.368455 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.368493 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:34Z","lastTransitionTime":"2025-10-02T09:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.470571 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.470642 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.470653 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.470668 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.470679 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:34Z","lastTransitionTime":"2025-10-02T09:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.572894 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.572930 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.572939 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.572973 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.572985 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:34Z","lastTransitionTime":"2025-10-02T09:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.675398 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.675443 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.675454 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.675478 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.675490 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:34Z","lastTransitionTime":"2025-10-02T09:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.778179 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.778224 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.778232 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.778249 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.778259 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:34Z","lastTransitionTime":"2025-10-02T09:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.881712 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.881785 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.881807 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.881834 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.881852 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:34Z","lastTransitionTime":"2025-10-02T09:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.990739 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.990798 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.990809 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.990838 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:34 crc kubenswrapper[4934]: I1002 09:50:34.990938 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:34Z","lastTransitionTime":"2025-10-02T09:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.093270 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.093524 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.093705 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.093778 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.093843 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:35Z","lastTransitionTime":"2025-10-02T09:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.196476 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.196907 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.196981 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.197116 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.197194 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:35Z","lastTransitionTime":"2025-10-02T09:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.300194 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.300471 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.300556 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.300684 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.300811 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:35Z","lastTransitionTime":"2025-10-02T09:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.403869 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.404292 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.404449 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.404540 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.404638 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:35Z","lastTransitionTime":"2025-10-02T09:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.508688 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.508761 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.508785 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.508814 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.508836 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:35Z","lastTransitionTime":"2025-10-02T09:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.612301 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.612369 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.612390 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.612411 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.612425 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:35Z","lastTransitionTime":"2025-10-02T09:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.715521 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.715628 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.715646 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.715667 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.715679 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:35Z","lastTransitionTime":"2025-10-02T09:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.818892 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.818948 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.818958 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.818977 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.818987 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:35Z","lastTransitionTime":"2025-10-02T09:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.912833 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.912936 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:50:35 crc kubenswrapper[4934]: E1002 09:50:35.912978 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.913047 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.913097 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:50:35 crc kubenswrapper[4934]: E1002 09:50:35.913139 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:50:35 crc kubenswrapper[4934]: E1002 09:50:35.913201 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:50:35 crc kubenswrapper[4934]: E1002 09:50:35.913275 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.922540 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.922636 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.922661 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.922687 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:35 crc kubenswrapper[4934]: I1002 09:50:35.922706 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:35Z","lastTransitionTime":"2025-10-02T09:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.024839 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.024879 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.024890 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.024904 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.024915 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:36Z","lastTransitionTime":"2025-10-02T09:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.128516 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.128668 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.128699 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.128730 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.128752 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:36Z","lastTransitionTime":"2025-10-02T09:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.231270 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.231362 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.231389 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.231419 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.231436 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:36Z","lastTransitionTime":"2025-10-02T09:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.334848 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.334899 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.334914 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.334932 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.334943 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:36Z","lastTransitionTime":"2025-10-02T09:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.437728 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.437765 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.437779 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.437793 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.437804 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:36Z","lastTransitionTime":"2025-10-02T09:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.540328 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.540363 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.540373 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.540385 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.540395 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:36Z","lastTransitionTime":"2025-10-02T09:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.643086 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.643129 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.643140 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.643157 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.643168 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:36Z","lastTransitionTime":"2025-10-02T09:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.746127 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.746184 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.746196 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.746212 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.746223 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:36Z","lastTransitionTime":"2025-10-02T09:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.848989 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.849032 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.849044 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.849062 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.849074 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:36Z","lastTransitionTime":"2025-10-02T09:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.951901 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.951944 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.951953 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.951972 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:36 crc kubenswrapper[4934]: I1002 09:50:36.951982 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:36Z","lastTransitionTime":"2025-10-02T09:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.055260 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.055334 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.055356 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.055384 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.055405 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:37Z","lastTransitionTime":"2025-10-02T09:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.158566 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.158664 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.158677 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.158695 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.158707 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:37Z","lastTransitionTime":"2025-10-02T09:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.216662 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.216717 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.216735 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.216759 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.216777 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:50:37Z","lastTransitionTime":"2025-10-02T09:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.268399 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-p9tq6"] Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.268858 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p9tq6" Oct 02 09:50:37 crc kubenswrapper[4934]: W1002 09:50:37.270616 4934 reflector.go:561] object-"openshift-cluster-version"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-cluster-version": no relationship found between node 'crc' and this object Oct 02 09:50:37 crc kubenswrapper[4934]: E1002 09:50:37.270663 4934 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-version\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-cluster-version\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 02 09:50:37 crc kubenswrapper[4934]: W1002 09:50:37.270803 4934 reflector.go:561] object-"openshift-cluster-version"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-cluster-version": no relationship found between node 'crc' and this object Oct 02 09:50:37 crc kubenswrapper[4934]: E1002 09:50:37.270854 4934 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-version\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-cluster-version\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 02 09:50:37 crc kubenswrapper[4934]: W1002 09:50:37.272518 4934 reflector.go:561] object-"openshift-cluster-version"/"default-dockercfg-gxtc4": failed to list *v1.Secret: secrets "default-dockercfg-gxtc4" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-cluster-version": no relationship found between node 'crc' and this object Oct 02 09:50:37 crc kubenswrapper[4934]: E1002 09:50:37.272620 4934 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-version\"/\"default-dockercfg-gxtc4\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"default-dockercfg-gxtc4\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-cluster-version\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 02 09:50:37 crc kubenswrapper[4934]: W1002 09:50:37.272522 4934 reflector.go:561] object-"openshift-cluster-version"/"cluster-version-operator-serving-cert": failed to list *v1.Secret: secrets "cluster-version-operator-serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-cluster-version": no relationship found between node 'crc' and this object Oct 02 09:50:37 crc kubenswrapper[4934]: E1002 09:50:37.272670 4934 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-version\"/\"cluster-version-operator-serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cluster-version-operator-serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-cluster-version\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.297874 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-79fxg" podStartSLOduration=88.297854467 podStartE2EDuration="1m28.297854467s" podCreationTimestamp="2025-10-02 09:49:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:37.287039117 +0000 UTC m=+109.039680649" watchObservedRunningTime="2025-10-02 09:50:37.297854467 +0000 UTC m=+109.050495989" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.299195 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/dbde12e9-5c31-44ee-b093-0a23c5b0b7d7-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-p9tq6\" (UID: \"dbde12e9-5c31-44ee-b093-0a23c5b0b7d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p9tq6" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.299241 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dbde12e9-5c31-44ee-b093-0a23c5b0b7d7-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-p9tq6\" (UID: \"dbde12e9-5c31-44ee-b093-0a23c5b0b7d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p9tq6" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.299283 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dbde12e9-5c31-44ee-b093-0a23c5b0b7d7-service-ca\") pod \"cluster-version-operator-5c965bbfc6-p9tq6\" (UID: \"dbde12e9-5c31-44ee-b093-0a23c5b0b7d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p9tq6" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.299331 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/dbde12e9-5c31-44ee-b093-0a23c5b0b7d7-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-p9tq6\" (UID: \"dbde12e9-5c31-44ee-b093-0a23c5b0b7d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p9tq6" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.299351 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dbde12e9-5c31-44ee-b093-0a23c5b0b7d7-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-p9tq6\" (UID: \"dbde12e9-5c31-44ee-b093-0a23c5b0b7d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p9tq6" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.377848 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-tctph" podStartSLOduration=88.377826483 podStartE2EDuration="1m28.377826483s" podCreationTimestamp="2025-10-02 09:49:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:37.377497026 +0000 UTC m=+109.130138558" watchObservedRunningTime="2025-10-02 09:50:37.377826483 +0000 UTC m=+109.130468005" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.390687 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=89.390670279 podStartE2EDuration="1m29.390670279s" podCreationTimestamp="2025-10-02 09:49:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:37.390406843 +0000 UTC m=+109.143048385" watchObservedRunningTime="2025-10-02 09:50:37.390670279 +0000 UTC m=+109.143311801" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.400006 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/dbde12e9-5c31-44ee-b093-0a23c5b0b7d7-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-p9tq6\" (UID: \"dbde12e9-5c31-44ee-b093-0a23c5b0b7d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p9tq6" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.400271 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dbde12e9-5c31-44ee-b093-0a23c5b0b7d7-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-p9tq6\" (UID: \"dbde12e9-5c31-44ee-b093-0a23c5b0b7d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p9tq6" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.400397 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dbde12e9-5c31-44ee-b093-0a23c5b0b7d7-service-ca\") pod \"cluster-version-operator-5c965bbfc6-p9tq6\" (UID: \"dbde12e9-5c31-44ee-b093-0a23c5b0b7d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p9tq6" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.400508 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/dbde12e9-5c31-44ee-b093-0a23c5b0b7d7-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-p9tq6\" (UID: \"dbde12e9-5c31-44ee-b093-0a23c5b0b7d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p9tq6" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.400644 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dbde12e9-5c31-44ee-b093-0a23c5b0b7d7-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-p9tq6\" (UID: \"dbde12e9-5c31-44ee-b093-0a23c5b0b7d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p9tq6" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.400547 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/dbde12e9-5c31-44ee-b093-0a23c5b0b7d7-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-p9tq6\" (UID: \"dbde12e9-5c31-44ee-b093-0a23c5b0b7d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p9tq6" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.400126 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/dbde12e9-5c31-44ee-b093-0a23c5b0b7d7-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-p9tq6\" (UID: \"dbde12e9-5c31-44ee-b093-0a23c5b0b7d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p9tq6" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.405364 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=52.405349368 podStartE2EDuration="52.405349368s" podCreationTimestamp="2025-10-02 09:49:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:37.405184925 +0000 UTC m=+109.157826447" watchObservedRunningTime="2025-10-02 09:50:37.405349368 +0000 UTC m=+109.157990890" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.419256 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-85mph" podStartSLOduration=88.419237679 podStartE2EDuration="1m28.419237679s" podCreationTimestamp="2025-10-02 09:49:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:37.418824429 +0000 UTC m=+109.171465951" watchObservedRunningTime="2025-10-02 09:50:37.419237679 +0000 UTC m=+109.171879201" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.451722 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podStartSLOduration=88.451701908 podStartE2EDuration="1m28.451701908s" podCreationTimestamp="2025-10-02 09:49:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:37.430394477 +0000 UTC m=+109.183036009" watchObservedRunningTime="2025-10-02 09:50:37.451701908 +0000 UTC m=+109.204343430" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.470645 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ffb9r" podStartSLOduration=87.470624525 podStartE2EDuration="1m27.470624525s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:37.470333788 +0000 UTC m=+109.222975330" watchObservedRunningTime="2025-10-02 09:50:37.470624525 +0000 UTC m=+109.223266047" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.486948 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=88.486932151 podStartE2EDuration="1m28.486932151s" podCreationTimestamp="2025-10-02 09:49:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:37.48604454 +0000 UTC m=+109.238686062" watchObservedRunningTime="2025-10-02 09:50:37.486932151 +0000 UTC m=+109.239573673" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.531905 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-ds45z" podStartSLOduration=87.531882858 podStartE2EDuration="1m27.531882858s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:37.519615595 +0000 UTC m=+109.272257137" watchObservedRunningTime="2025-10-02 09:50:37.531882858 +0000 UTC m=+109.284524380" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.544679 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=25.544661893 podStartE2EDuration="25.544661893s" podCreationTimestamp="2025-10-02 09:50:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:37.543448485 +0000 UTC m=+109.296090017" watchObservedRunningTime="2025-10-02 09:50:37.544661893 +0000 UTC m=+109.297303415" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.913074 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:50:37 crc kubenswrapper[4934]: E1002 09:50:37.913197 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.913228 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.913085 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:50:37 crc kubenswrapper[4934]: I1002 09:50:37.913290 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:50:37 crc kubenswrapper[4934]: E1002 09:50:37.913319 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:50:37 crc kubenswrapper[4934]: E1002 09:50:37.913417 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:50:37 crc kubenswrapper[4934]: E1002 09:50:37.913624 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:50:38 crc kubenswrapper[4934]: I1002 09:50:38.171035 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 02 09:50:38 crc kubenswrapper[4934]: I1002 09:50:38.192273 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 02 09:50:38 crc kubenswrapper[4934]: I1002 09:50:38.203719 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dbde12e9-5c31-44ee-b093-0a23c5b0b7d7-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-p9tq6\" (UID: \"dbde12e9-5c31-44ee-b093-0a23c5b0b7d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p9tq6" Oct 02 09:50:38 crc kubenswrapper[4934]: E1002 09:50:38.401631 4934 configmap.go:193] Couldn't get configMap openshift-cluster-version/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Oct 02 09:50:38 crc kubenswrapper[4934]: E1002 09:50:38.401729 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/dbde12e9-5c31-44ee-b093-0a23c5b0b7d7-service-ca podName:dbde12e9-5c31-44ee-b093-0a23c5b0b7d7 nodeName:}" failed. No retries permitted until 2025-10-02 09:50:38.901708172 +0000 UTC m=+110.654349704 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "service-ca" (UniqueName: "kubernetes.io/configmap/dbde12e9-5c31-44ee-b093-0a23c5b0b7d7-service-ca") pod "cluster-version-operator-5c965bbfc6-p9tq6" (UID: "dbde12e9-5c31-44ee-b093-0a23c5b0b7d7") : failed to sync configmap cache: timed out waiting for the condition Oct 02 09:50:38 crc kubenswrapper[4934]: E1002 09:50:38.401635 4934 secret.go:188] Couldn't get secret openshift-cluster-version/cluster-version-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 02 09:50:38 crc kubenswrapper[4934]: E1002 09:50:38.401848 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dbde12e9-5c31-44ee-b093-0a23c5b0b7d7-serving-cert podName:dbde12e9-5c31-44ee-b093-0a23c5b0b7d7 nodeName:}" failed. No retries permitted until 2025-10-02 09:50:38.901824395 +0000 UTC m=+110.654465917 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/dbde12e9-5c31-44ee-b093-0a23c5b0b7d7-serving-cert") pod "cluster-version-operator-5c965bbfc6-p9tq6" (UID: "dbde12e9-5c31-44ee-b093-0a23c5b0b7d7") : failed to sync secret cache: timed out waiting for the condition Oct 02 09:50:38 crc kubenswrapper[4934]: I1002 09:50:38.673390 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 02 09:50:38 crc kubenswrapper[4934]: I1002 09:50:38.871512 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 02 09:50:38 crc kubenswrapper[4934]: I1002 09:50:38.919384 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dbde12e9-5c31-44ee-b093-0a23c5b0b7d7-service-ca\") pod \"cluster-version-operator-5c965bbfc6-p9tq6\" (UID: \"dbde12e9-5c31-44ee-b093-0a23c5b0b7d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p9tq6" Oct 02 09:50:38 crc kubenswrapper[4934]: I1002 09:50:38.919437 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dbde12e9-5c31-44ee-b093-0a23c5b0b7d7-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-p9tq6\" (UID: \"dbde12e9-5c31-44ee-b093-0a23c5b0b7d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p9tq6" Oct 02 09:50:38 crc kubenswrapper[4934]: I1002 09:50:38.920531 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dbde12e9-5c31-44ee-b093-0a23c5b0b7d7-service-ca\") pod \"cluster-version-operator-5c965bbfc6-p9tq6\" (UID: \"dbde12e9-5c31-44ee-b093-0a23c5b0b7d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p9tq6" Oct 02 09:50:38 crc kubenswrapper[4934]: I1002 09:50:38.924691 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dbde12e9-5c31-44ee-b093-0a23c5b0b7d7-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-p9tq6\" (UID: \"dbde12e9-5c31-44ee-b093-0a23c5b0b7d7\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p9tq6" Oct 02 09:50:39 crc kubenswrapper[4934]: I1002 09:50:39.091495 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p9tq6" Oct 02 09:50:39 crc kubenswrapper[4934]: I1002 09:50:39.470365 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p9tq6" event={"ID":"dbde12e9-5c31-44ee-b093-0a23c5b0b7d7","Type":"ContainerStarted","Data":"b47257c84a289f10e866352aa7ee7bbc095f1419dcdca84fb66c631080874f8d"} Oct 02 09:50:39 crc kubenswrapper[4934]: I1002 09:50:39.470418 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p9tq6" event={"ID":"dbde12e9-5c31-44ee-b093-0a23c5b0b7d7","Type":"ContainerStarted","Data":"03c58f30731651d1a10a3a03d0efd60c7af24da9580ba677b0ea982cd6282d26"} Oct 02 09:50:39 crc kubenswrapper[4934]: I1002 09:50:39.486368 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=89.486351903 podStartE2EDuration="1m29.486351903s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:37.568144825 +0000 UTC m=+109.320786347" watchObservedRunningTime="2025-10-02 09:50:39.486351903 +0000 UTC m=+111.238993425" Oct 02 09:50:39 crc kubenswrapper[4934]: I1002 09:50:39.486610 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p9tq6" podStartSLOduration=90.486605859 podStartE2EDuration="1m30.486605859s" podCreationTimestamp="2025-10-02 09:49:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:39.486263251 +0000 UTC m=+111.238904773" watchObservedRunningTime="2025-10-02 09:50:39.486605859 +0000 UTC m=+111.239247381" Oct 02 09:50:39 crc kubenswrapper[4934]: I1002 09:50:39.912500 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:50:39 crc kubenswrapper[4934]: I1002 09:50:39.912621 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:50:39 crc kubenswrapper[4934]: E1002 09:50:39.912655 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:50:39 crc kubenswrapper[4934]: E1002 09:50:39.912748 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:50:39 crc kubenswrapper[4934]: I1002 09:50:39.912500 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:50:39 crc kubenswrapper[4934]: E1002 09:50:39.912833 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:50:39 crc kubenswrapper[4934]: I1002 09:50:39.913274 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:50:39 crc kubenswrapper[4934]: E1002 09:50:39.913558 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:50:41 crc kubenswrapper[4934]: I1002 09:50:41.912639 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:50:41 crc kubenswrapper[4934]: I1002 09:50:41.912680 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:50:41 crc kubenswrapper[4934]: I1002 09:50:41.912680 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:50:41 crc kubenswrapper[4934]: I1002 09:50:41.912783 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:50:41 crc kubenswrapper[4934]: E1002 09:50:41.912773 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:50:41 crc kubenswrapper[4934]: E1002 09:50:41.912891 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:50:41 crc kubenswrapper[4934]: I1002 09:50:41.913573 4934 scope.go:117] "RemoveContainer" containerID="358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe" Oct 02 09:50:41 crc kubenswrapper[4934]: E1002 09:50:41.913799 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:50:41 crc kubenswrapper[4934]: E1002 09:50:41.914064 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:50:42 crc kubenswrapper[4934]: I1002 09:50:42.480481 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-th4v6_1d4841b5-0469-461e-875c-25b9fe848141/ovnkube-controller/3.log" Oct 02 09:50:42 crc kubenswrapper[4934]: I1002 09:50:42.483533 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" event={"ID":"1d4841b5-0469-461e-875c-25b9fe848141","Type":"ContainerStarted","Data":"354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66"} Oct 02 09:50:42 crc kubenswrapper[4934]: I1002 09:50:42.483987 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:50:42 crc kubenswrapper[4934]: I1002 09:50:42.520796 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" podStartSLOduration=92.520779591 podStartE2EDuration="1m32.520779591s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:42.520681239 +0000 UTC m=+114.273322761" watchObservedRunningTime="2025-10-02 09:50:42.520779591 +0000 UTC m=+114.273421113" Oct 02 09:50:42 crc kubenswrapper[4934]: I1002 09:50:42.729114 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-n7cgz"] Oct 02 09:50:42 crc kubenswrapper[4934]: I1002 09:50:42.729245 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:50:42 crc kubenswrapper[4934]: E1002 09:50:42.729350 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:50:43 crc kubenswrapper[4934]: I1002 09:50:43.912670 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:50:43 crc kubenswrapper[4934]: I1002 09:50:43.912710 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:50:43 crc kubenswrapper[4934]: I1002 09:50:43.912772 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:50:43 crc kubenswrapper[4934]: E1002 09:50:43.912854 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:50:43 crc kubenswrapper[4934]: E1002 09:50:43.913023 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:50:43 crc kubenswrapper[4934]: E1002 09:50:43.913072 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:50:44 crc kubenswrapper[4934]: I1002 09:50:44.912598 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:50:44 crc kubenswrapper[4934]: E1002 09:50:44.912745 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7cgz" podUID="0cbef3ee-c66f-47fa-94d3-5ce9892c403f" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.193203 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.193402 4934 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.242848 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.243662 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-fc4l5"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.243985 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.244317 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.244696 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gmxzf"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.245327 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q2p9l"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.245704 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-q2p9l" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.245958 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.246242 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gmxzf" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.246377 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.249685 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-t6dmx"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.250396 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-t6dmx" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.251679 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.251851 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.252060 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.252100 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.252216 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.252227 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.252654 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-f44r4"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.252726 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.252892 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.253047 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.253250 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.253371 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-f44r4" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.253718 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.253904 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.255329 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxlv9"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.255485 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.255829 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.255937 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxlv9" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.255984 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.256145 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.256497 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7t8rp"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.257181 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7t8rp" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.258174 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-vhxn6"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.259093 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rjhkx"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.260077 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6zlqv"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.260413 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vhxn6" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.260971 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-rjhkx" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.261454 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6zlqv" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.267707 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-mcnd2"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.268594 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-mcnd2" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.268719 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4pp8p"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.269239 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.271301 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-4bqtk"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.284685 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4bqtk" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.291929 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-5wrws"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.300588 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-5wrws" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.300870 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jht5p"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.301544 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.301645 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d074a40-447d-4144-a2dc-ca69269c84de-serving-cert\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.301674 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/62e9cfe8-c134-423b-b971-c552f98526f9-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-gmxzf\" (UID: \"62e9cfe8-c134-423b-b971-c552f98526f9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gmxzf" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.301707 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8bbc993a-a1ea-48ae-a765-8f656d17329c-images\") pod \"machine-api-operator-5694c8668f-t6dmx\" (UID: \"8bbc993a-a1ea-48ae-a765-8f656d17329c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t6dmx" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.301737 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1d074a40-447d-4144-a2dc-ca69269c84de-image-import-ca\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.301757 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1d074a40-447d-4144-a2dc-ca69269c84de-encryption-config\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.301773 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-client-ca\") pod \"controller-manager-879f6c89f-q2p9l\" (UID: \"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q2p9l" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.301791 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6096bf5b-7d9b-41c1-a6b9-d2f6879175a6-audit-policies\") pod \"apiserver-7bbb656c7d-lfhf5\" (UID: \"6096bf5b-7d9b-41c1-a6b9-d2f6879175a6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.301808 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6096bf5b-7d9b-41c1-a6b9-d2f6879175a6-etcd-client\") pod \"apiserver-7bbb656c7d-lfhf5\" (UID: \"6096bf5b-7d9b-41c1-a6b9-d2f6879175a6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.301824 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1d074a40-447d-4144-a2dc-ca69269c84de-trusted-ca-bundle\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.301839 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/87b8c00f-e910-42f0-8089-73df46c6cf0b-client-ca\") pod \"route-controller-manager-6576b87f9c-d6cv6\" (UID: \"87b8c00f-e910-42f0-8089-73df46c6cf0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.301859 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/1d074a40-447d-4144-a2dc-ca69269c84de-node-pullsecrets\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.301890 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4d8hf\" (UniqueName: \"kubernetes.io/projected/1d074a40-447d-4144-a2dc-ca69269c84de-kube-api-access-4d8hf\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.301906 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svdvd\" (UniqueName: \"kubernetes.io/projected/87b8c00f-e910-42f0-8089-73df46c6cf0b-kube-api-access-svdvd\") pod \"route-controller-manager-6576b87f9c-d6cv6\" (UID: \"87b8c00f-e910-42f0-8089-73df46c6cf0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.301925 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-config\") pod \"controller-manager-879f6c89f-q2p9l\" (UID: \"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q2p9l" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.301953 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rchq\" (UniqueName: \"kubernetes.io/projected/8bbc993a-a1ea-48ae-a765-8f656d17329c-kube-api-access-9rchq\") pod \"machine-api-operator-5694c8668f-t6dmx\" (UID: \"8bbc993a-a1ea-48ae-a765-8f656d17329c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t6dmx" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.301969 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1d074a40-447d-4144-a2dc-ca69269c84de-audit\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.301988 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn9w7\" (UniqueName: \"kubernetes.io/projected/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-kube-api-access-zn9w7\") pod \"controller-manager-879f6c89f-q2p9l\" (UID: \"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q2p9l" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.302006 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6096bf5b-7d9b-41c1-a6b9-d2f6879175a6-encryption-config\") pod \"apiserver-7bbb656c7d-lfhf5\" (UID: \"6096bf5b-7d9b-41c1-a6b9-d2f6879175a6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.302024 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/8bbc993a-a1ea-48ae-a765-8f656d17329c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-t6dmx\" (UID: \"8bbc993a-a1ea-48ae-a765-8f656d17329c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t6dmx" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.302043 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1d074a40-447d-4144-a2dc-ca69269c84de-etcd-client\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.302058 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6096bf5b-7d9b-41c1-a6b9-d2f6879175a6-serving-cert\") pod \"apiserver-7bbb656c7d-lfhf5\" (UID: \"6096bf5b-7d9b-41c1-a6b9-d2f6879175a6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.302075 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hz4wv\" (UniqueName: \"kubernetes.io/projected/6096bf5b-7d9b-41c1-a6b9-d2f6879175a6-kube-api-access-hz4wv\") pod \"apiserver-7bbb656c7d-lfhf5\" (UID: \"6096bf5b-7d9b-41c1-a6b9-d2f6879175a6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.302090 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-serving-cert\") pod \"controller-manager-879f6c89f-q2p9l\" (UID: \"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q2p9l" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.302108 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6096bf5b-7d9b-41c1-a6b9-d2f6879175a6-audit-dir\") pod \"apiserver-7bbb656c7d-lfhf5\" (UID: \"6096bf5b-7d9b-41c1-a6b9-d2f6879175a6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.302127 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bbc993a-a1ea-48ae-a765-8f656d17329c-config\") pod \"machine-api-operator-5694c8668f-t6dmx\" (UID: \"8bbc993a-a1ea-48ae-a765-8f656d17329c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t6dmx" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.302153 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d074a40-447d-4144-a2dc-ca69269c84de-config\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.302170 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-q2p9l\" (UID: \"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q2p9l" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.302184 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1d074a40-447d-4144-a2dc-ca69269c84de-audit-dir\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.302202 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87b8c00f-e910-42f0-8089-73df46c6cf0b-config\") pod \"route-controller-manager-6576b87f9c-d6cv6\" (UID: \"87b8c00f-e910-42f0-8089-73df46c6cf0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.302227 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6096bf5b-7d9b-41c1-a6b9-d2f6879175a6-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-lfhf5\" (UID: \"6096bf5b-7d9b-41c1-a6b9-d2f6879175a6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.302250 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87b8c00f-e910-42f0-8089-73df46c6cf0b-serving-cert\") pod \"route-controller-manager-6576b87f9c-d6cv6\" (UID: \"87b8c00f-e910-42f0-8089-73df46c6cf0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.302274 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6096bf5b-7d9b-41c1-a6b9-d2f6879175a6-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-lfhf5\" (UID: \"6096bf5b-7d9b-41c1-a6b9-d2f6879175a6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.302302 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1d074a40-447d-4144-a2dc-ca69269c84de-etcd-serving-ca\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.302319 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9j26x\" (UniqueName: \"kubernetes.io/projected/62e9cfe8-c134-423b-b971-c552f98526f9-kube-api-access-9j26x\") pod \"cluster-samples-operator-665b6dd947-gmxzf\" (UID: \"62e9cfe8-c134-423b-b971-c552f98526f9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gmxzf" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.304279 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qg98s"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.307126 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.307163 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.307266 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.307302 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.307372 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.307408 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.307490 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.307512 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.307538 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.307594 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.307656 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.307742 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.307786 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.307833 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.307936 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.308027 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.308133 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.308181 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.308215 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.308257 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.308330 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.308406 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.308467 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.308545 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.308624 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.308644 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.308716 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.308723 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.308813 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.308821 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.308871 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qg98s" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.308904 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.308913 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.308953 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.308989 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.309018 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.309074 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.309093 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.308990 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.309171 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.309236 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.309239 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.309297 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.309325 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.309375 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.309447 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.309519 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.309626 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.309708 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.309170 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.309821 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.309832 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.309891 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.309926 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.309998 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.310037 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.310103 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.310135 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.310184 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.310376 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.310820 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.311287 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb67r"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.311763 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb67r" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.314553 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.314699 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.314938 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.314946 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.316311 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.316511 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.316649 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.316793 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.318107 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.318200 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-wjz5g"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.318241 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.318523 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.318975 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7qgvq"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.319123 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.319357 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.319415 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7qgvq" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.319480 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.319581 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.319722 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.319759 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-wjz5g" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.319859 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.320190 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bnvc4"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.320531 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bnvc4" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.320671 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.320743 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-66hrr"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.321345 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-66hrr" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.345829 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.362928 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.363151 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.364726 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.368662 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-x7mrm"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.369545 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-86wbg"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.369873 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-7frl9"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.370409 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7frl9" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.370773 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-x7mrm" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.371071 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-86wbg" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.371381 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.378175 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9ts9g"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.378878 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5s688"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.379399 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323305-jv64v"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.379835 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-jv64v" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.380357 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9ts9g" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.380535 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5s688" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.381932 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.385813 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-6jtqh"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.386511 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.386815 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hw97b"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.387076 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6jtqh" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.387358 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hw97b" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.389216 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-hr945"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.390372 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lch8r"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.391009 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-pzrnw"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.391070 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lch8r" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.391469 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-hr945" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.391501 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-pzrnw" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.392852 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8xjjw"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.393904 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-llqjv"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.393956 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8xjjw" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.394734 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-llqjv" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.395170 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z2d6w"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.395808 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z2d6w" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.397709 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wlf9l"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.398528 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-2vn7r"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.399076 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-2vn7r" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.399314 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wlf9l" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.403720 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6096bf5b-7d9b-41c1-a6b9-d2f6879175a6-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-lfhf5\" (UID: \"6096bf5b-7d9b-41c1-a6b9-d2f6879175a6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.403759 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1d074a40-447d-4144-a2dc-ca69269c84de-etcd-serving-ca\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.403781 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/602dc2aa-bffc-4e88-95ad-5dce0fc6d6df-config\") pod \"machine-approver-56656f9798-4bqtk\" (UID: \"602dc2aa-bffc-4e88-95ad-5dce0fc6d6df\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4bqtk" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.403799 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9j26x\" (UniqueName: \"kubernetes.io/projected/62e9cfe8-c134-423b-b971-c552f98526f9-kube-api-access-9j26x\") pod \"cluster-samples-operator-665b6dd947-gmxzf\" (UID: \"62e9cfe8-c134-423b-b971-c552f98526f9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gmxzf" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.403821 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/62e9cfe8-c134-423b-b971-c552f98526f9-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-gmxzf\" (UID: \"62e9cfe8-c134-423b-b971-c552f98526f9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gmxzf" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.403838 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8bbc993a-a1ea-48ae-a765-8f656d17329c-images\") pod \"machine-api-operator-5694c8668f-t6dmx\" (UID: \"8bbc993a-a1ea-48ae-a765-8f656d17329c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t6dmx" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.403855 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76e57d57-21b3-4f79-a7ee-b6bf6117c8d3-serving-cert\") pod \"openshift-config-operator-7777fb866f-vhxn6\" (UID: \"76e57d57-21b3-4f79-a7ee-b6bf6117c8d3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vhxn6" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.403871 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/602dc2aa-bffc-4e88-95ad-5dce0fc6d6df-machine-approver-tls\") pod \"machine-approver-56656f9798-4bqtk\" (UID: \"602dc2aa-bffc-4e88-95ad-5dce0fc6d6df\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4bqtk" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.403887 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d074a40-447d-4144-a2dc-ca69269c84de-serving-cert\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.403909 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1d074a40-447d-4144-a2dc-ca69269c84de-image-import-ca\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.403928 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1d074a40-447d-4144-a2dc-ca69269c84de-encryption-config\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.403947 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-client-ca\") pod \"controller-manager-879f6c89f-q2p9l\" (UID: \"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q2p9l" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.403962 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6096bf5b-7d9b-41c1-a6b9-d2f6879175a6-audit-policies\") pod \"apiserver-7bbb656c7d-lfhf5\" (UID: \"6096bf5b-7d9b-41c1-a6b9-d2f6879175a6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.403977 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jplkg\" (UniqueName: \"kubernetes.io/projected/602dc2aa-bffc-4e88-95ad-5dce0fc6d6df-kube-api-access-jplkg\") pod \"machine-approver-56656f9798-4bqtk\" (UID: \"602dc2aa-bffc-4e88-95ad-5dce0fc6d6df\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4bqtk" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.403992 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6096bf5b-7d9b-41c1-a6b9-d2f6879175a6-etcd-client\") pod \"apiserver-7bbb656c7d-lfhf5\" (UID: \"6096bf5b-7d9b-41c1-a6b9-d2f6879175a6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.404007 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1d074a40-447d-4144-a2dc-ca69269c84de-trusted-ca-bundle\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.404027 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/87b8c00f-e910-42f0-8089-73df46c6cf0b-client-ca\") pod \"route-controller-manager-6576b87f9c-d6cv6\" (UID: \"87b8c00f-e910-42f0-8089-73df46c6cf0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.404053 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/1d074a40-447d-4144-a2dc-ca69269c84de-node-pullsecrets\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.404422 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/602dc2aa-bffc-4e88-95ad-5dce0fc6d6df-auth-proxy-config\") pod \"machine-approver-56656f9798-4bqtk\" (UID: \"602dc2aa-bffc-4e88-95ad-5dce0fc6d6df\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4bqtk" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.404449 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4d8hf\" (UniqueName: \"kubernetes.io/projected/1d074a40-447d-4144-a2dc-ca69269c84de-kube-api-access-4d8hf\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.404467 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svdvd\" (UniqueName: \"kubernetes.io/projected/87b8c00f-e910-42f0-8089-73df46c6cf0b-kube-api-access-svdvd\") pod \"route-controller-manager-6576b87f9c-d6cv6\" (UID: \"87b8c00f-e910-42f0-8089-73df46c6cf0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.404482 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-config\") pod \"controller-manager-879f6c89f-q2p9l\" (UID: \"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q2p9l" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.404497 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1d074a40-447d-4144-a2dc-ca69269c84de-audit\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.404514 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn9w7\" (UniqueName: \"kubernetes.io/projected/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-kube-api-access-zn9w7\") pod \"controller-manager-879f6c89f-q2p9l\" (UID: \"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q2p9l" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.404528 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rchq\" (UniqueName: \"kubernetes.io/projected/8bbc993a-a1ea-48ae-a765-8f656d17329c-kube-api-access-9rchq\") pod \"machine-api-operator-5694c8668f-t6dmx\" (UID: \"8bbc993a-a1ea-48ae-a765-8f656d17329c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t6dmx" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.404545 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fwxz\" (UniqueName: \"kubernetes.io/projected/76e57d57-21b3-4f79-a7ee-b6bf6117c8d3-kube-api-access-2fwxz\") pod \"openshift-config-operator-7777fb866f-vhxn6\" (UID: \"76e57d57-21b3-4f79-a7ee-b6bf6117c8d3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vhxn6" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.404566 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6096bf5b-7d9b-41c1-a6b9-d2f6879175a6-encryption-config\") pod \"apiserver-7bbb656c7d-lfhf5\" (UID: \"6096bf5b-7d9b-41c1-a6b9-d2f6879175a6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.404584 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/8bbc993a-a1ea-48ae-a765-8f656d17329c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-t6dmx\" (UID: \"8bbc993a-a1ea-48ae-a765-8f656d17329c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t6dmx" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.404614 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1d074a40-447d-4144-a2dc-ca69269c84de-etcd-client\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.404632 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-serving-cert\") pod \"controller-manager-879f6c89f-q2p9l\" (UID: \"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q2p9l" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.404647 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6096bf5b-7d9b-41c1-a6b9-d2f6879175a6-serving-cert\") pod \"apiserver-7bbb656c7d-lfhf5\" (UID: \"6096bf5b-7d9b-41c1-a6b9-d2f6879175a6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.404665 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hz4wv\" (UniqueName: \"kubernetes.io/projected/6096bf5b-7d9b-41c1-a6b9-d2f6879175a6-kube-api-access-hz4wv\") pod \"apiserver-7bbb656c7d-lfhf5\" (UID: \"6096bf5b-7d9b-41c1-a6b9-d2f6879175a6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.404682 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6096bf5b-7d9b-41c1-a6b9-d2f6879175a6-audit-dir\") pod \"apiserver-7bbb656c7d-lfhf5\" (UID: \"6096bf5b-7d9b-41c1-a6b9-d2f6879175a6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.404698 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bbc993a-a1ea-48ae-a765-8f656d17329c-config\") pod \"machine-api-operator-5694c8668f-t6dmx\" (UID: \"8bbc993a-a1ea-48ae-a765-8f656d17329c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t6dmx" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.404724 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d074a40-447d-4144-a2dc-ca69269c84de-config\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.404739 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-q2p9l\" (UID: \"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q2p9l" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.404761 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1d074a40-447d-4144-a2dc-ca69269c84de-audit-dir\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.404777 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87b8c00f-e910-42f0-8089-73df46c6cf0b-config\") pod \"route-controller-manager-6576b87f9c-d6cv6\" (UID: \"87b8c00f-e910-42f0-8089-73df46c6cf0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.404795 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6096bf5b-7d9b-41c1-a6b9-d2f6879175a6-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-lfhf5\" (UID: \"6096bf5b-7d9b-41c1-a6b9-d2f6879175a6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.404811 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87b8c00f-e910-42f0-8089-73df46c6cf0b-serving-cert\") pod \"route-controller-manager-6576b87f9c-d6cv6\" (UID: \"87b8c00f-e910-42f0-8089-73df46c6cf0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.404830 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/76e57d57-21b3-4f79-a7ee-b6bf6117c8d3-available-featuregates\") pod \"openshift-config-operator-7777fb866f-vhxn6\" (UID: \"76e57d57-21b3-4f79-a7ee-b6bf6117c8d3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vhxn6" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.405647 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gmxzf"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.405742 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-fc4l5"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.405763 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxlv9"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.405776 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bnvc4"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.405859 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-t6dmx"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.407052 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6096bf5b-7d9b-41c1-a6b9-d2f6879175a6-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-lfhf5\" (UID: \"6096bf5b-7d9b-41c1-a6b9-d2f6879175a6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.411614 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.413015 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8bbc993a-a1ea-48ae-a765-8f656d17329c-images\") pod \"machine-api-operator-5694c8668f-t6dmx\" (UID: \"8bbc993a-a1ea-48ae-a765-8f656d17329c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t6dmx" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.414340 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bbc993a-a1ea-48ae-a765-8f656d17329c-config\") pod \"machine-api-operator-5694c8668f-t6dmx\" (UID: \"8bbc993a-a1ea-48ae-a765-8f656d17329c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t6dmx" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.418195 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.418725 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.419020 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1d074a40-447d-4144-a2dc-ca69269c84de-audit-dir\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.419248 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1d074a40-447d-4144-a2dc-ca69269c84de-trusted-ca-bundle\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.419914 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d074a40-447d-4144-a2dc-ca69269c84de-config\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.420098 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/87b8c00f-e910-42f0-8089-73df46c6cf0b-client-ca\") pod \"route-controller-manager-6576b87f9c-d6cv6\" (UID: \"87b8c00f-e910-42f0-8089-73df46c6cf0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.420159 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/1d074a40-447d-4144-a2dc-ca69269c84de-node-pullsecrets\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.420443 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb67r"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.420501 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q2p9l"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.421348 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-config\") pod \"controller-manager-879f6c89f-q2p9l\" (UID: \"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q2p9l" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.421415 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-client-ca\") pod \"controller-manager-879f6c89f-q2p9l\" (UID: \"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q2p9l" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.421516 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1d074a40-447d-4144-a2dc-ca69269c84de-image-import-ca\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.421522 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-q2p9l\" (UID: \"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q2p9l" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.421789 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1d074a40-447d-4144-a2dc-ca69269c84de-audit\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.430743 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/62e9cfe8-c134-423b-b971-c552f98526f9-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-gmxzf\" (UID: \"62e9cfe8-c134-423b-b971-c552f98526f9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gmxzf" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.431066 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6096bf5b-7d9b-41c1-a6b9-d2f6879175a6-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-lfhf5\" (UID: \"6096bf5b-7d9b-41c1-a6b9-d2f6879175a6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.433264 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1d074a40-447d-4144-a2dc-ca69269c84de-etcd-client\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.433568 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6096bf5b-7d9b-41c1-a6b9-d2f6879175a6-audit-dir\") pod \"apiserver-7bbb656c7d-lfhf5\" (UID: \"6096bf5b-7d9b-41c1-a6b9-d2f6879175a6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.434821 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.435502 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1d074a40-447d-4144-a2dc-ca69269c84de-etcd-serving-ca\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.435508 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.435924 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6096bf5b-7d9b-41c1-a6b9-d2f6879175a6-audit-policies\") pod \"apiserver-7bbb656c7d-lfhf5\" (UID: \"6096bf5b-7d9b-41c1-a6b9-d2f6879175a6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.436104 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6096bf5b-7d9b-41c1-a6b9-d2f6879175a6-etcd-client\") pod \"apiserver-7bbb656c7d-lfhf5\" (UID: \"6096bf5b-7d9b-41c1-a6b9-d2f6879175a6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.436213 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6096bf5b-7d9b-41c1-a6b9-d2f6879175a6-serving-cert\") pod \"apiserver-7bbb656c7d-lfhf5\" (UID: \"6096bf5b-7d9b-41c1-a6b9-d2f6879175a6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.440348 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1d074a40-447d-4144-a2dc-ca69269c84de-encryption-config\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.441396 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7t8rp"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.443294 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87b8c00f-e910-42f0-8089-73df46c6cf0b-serving-cert\") pod \"route-controller-manager-6576b87f9c-d6cv6\" (UID: \"87b8c00f-e910-42f0-8089-73df46c6cf0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.443649 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.445863 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.449080 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d074a40-447d-4144-a2dc-ca69269c84de-serving-cert\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.453199 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/8bbc993a-a1ea-48ae-a765-8f656d17329c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-t6dmx\" (UID: \"8bbc993a-a1ea-48ae-a765-8f656d17329c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t6dmx" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.454756 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87b8c00f-e910-42f0-8089-73df46c6cf0b-config\") pod \"route-controller-manager-6576b87f9c-d6cv6\" (UID: \"87b8c00f-e910-42f0-8089-73df46c6cf0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.455126 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.456767 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.456915 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6096bf5b-7d9b-41c1-a6b9-d2f6879175a6-encryption-config\") pod \"apiserver-7bbb656c7d-lfhf5\" (UID: \"6096bf5b-7d9b-41c1-a6b9-d2f6879175a6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.458629 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-mcnd2"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.462824 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qg98s"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.464957 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-serving-cert\") pod \"controller-manager-879f6c89f-q2p9l\" (UID: \"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q2p9l" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.469776 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rjhkx"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.470107 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.474430 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-vhxn6"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.476653 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-86wbg"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.476705 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jht5p"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.478742 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hw97b"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.480182 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8xjjw"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.480715 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7qgvq"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.483053 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5s688"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.486236 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.487562 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wlf9l"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.488933 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-5wrws"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.489755 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-pzrnw"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.491266 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6zlqv"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.491681 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-2vn7r"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.492753 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-66hrr"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.493859 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4pp8p"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.497018 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9ts9g"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.500746 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-f44r4"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.502040 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-7frl9"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.504151 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323305-jv64v"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.504932 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lch8r"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.505741 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fwxz\" (UniqueName: \"kubernetes.io/projected/76e57d57-21b3-4f79-a7ee-b6bf6117c8d3-kube-api-access-2fwxz\") pod \"openshift-config-operator-7777fb866f-vhxn6\" (UID: \"76e57d57-21b3-4f79-a7ee-b6bf6117c8d3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vhxn6" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.505843 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/76e57d57-21b3-4f79-a7ee-b6bf6117c8d3-available-featuregates\") pod \"openshift-config-operator-7777fb866f-vhxn6\" (UID: \"76e57d57-21b3-4f79-a7ee-b6bf6117c8d3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vhxn6" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.505881 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/602dc2aa-bffc-4e88-95ad-5dce0fc6d6df-config\") pod \"machine-approver-56656f9798-4bqtk\" (UID: \"602dc2aa-bffc-4e88-95ad-5dce0fc6d6df\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4bqtk" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.505918 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76e57d57-21b3-4f79-a7ee-b6bf6117c8d3-serving-cert\") pod \"openshift-config-operator-7777fb866f-vhxn6\" (UID: \"76e57d57-21b3-4f79-a7ee-b6bf6117c8d3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vhxn6" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.505940 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/602dc2aa-bffc-4e88-95ad-5dce0fc6d6df-machine-approver-tls\") pod \"machine-approver-56656f9798-4bqtk\" (UID: \"602dc2aa-bffc-4e88-95ad-5dce0fc6d6df\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4bqtk" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.505977 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jplkg\" (UniqueName: \"kubernetes.io/projected/602dc2aa-bffc-4e88-95ad-5dce0fc6d6df-kube-api-access-jplkg\") pod \"machine-approver-56656f9798-4bqtk\" (UID: \"602dc2aa-bffc-4e88-95ad-5dce0fc6d6df\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4bqtk" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.506028 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/602dc2aa-bffc-4e88-95ad-5dce0fc6d6df-auth-proxy-config\") pod \"machine-approver-56656f9798-4bqtk\" (UID: \"602dc2aa-bffc-4e88-95ad-5dce0fc6d6df\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4bqtk" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.506220 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.507274 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/602dc2aa-bffc-4e88-95ad-5dce0fc6d6df-auth-proxy-config\") pod \"machine-approver-56656f9798-4bqtk\" (UID: \"602dc2aa-bffc-4e88-95ad-5dce0fc6d6df\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4bqtk" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.507328 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/602dc2aa-bffc-4e88-95ad-5dce0fc6d6df-config\") pod \"machine-approver-56656f9798-4bqtk\" (UID: \"602dc2aa-bffc-4e88-95ad-5dce0fc6d6df\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4bqtk" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.507503 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/76e57d57-21b3-4f79-a7ee-b6bf6117c8d3-available-featuregates\") pod \"openshift-config-operator-7777fb866f-vhxn6\" (UID: \"76e57d57-21b3-4f79-a7ee-b6bf6117c8d3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vhxn6" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.508976 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-hr945"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.509292 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/602dc2aa-bffc-4e88-95ad-5dce0fc6d6df-machine-approver-tls\") pod \"machine-approver-56656f9798-4bqtk\" (UID: \"602dc2aa-bffc-4e88-95ad-5dce0fc6d6df\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4bqtk" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.510412 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z2d6w"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.511620 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76e57d57-21b3-4f79-a7ee-b6bf6117c8d3-serving-cert\") pod \"openshift-config-operator-7777fb866f-vhxn6\" (UID: \"76e57d57-21b3-4f79-a7ee-b6bf6117c8d3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vhxn6" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.513365 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-6jtqh"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.514351 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-llqjv"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.515714 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-x7mrm"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.518192 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-w6z57"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.518762 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-w6z57" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.520436 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-v58f6"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.521000 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-v58f6" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.521491 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-v58f6"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.525972 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.546690 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.562511 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-pfl7n"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.563545 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-pfl7n" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.567515 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.575049 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-pfl7n"] Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.586486 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.608814 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.626850 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.647524 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.666756 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.686935 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.707263 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.726863 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.746678 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.766697 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.787102 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.806310 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.827429 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.846499 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.866651 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.887623 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.906842 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.913065 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.913093 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.913071 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.926752 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.955839 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 02 09:50:45 crc kubenswrapper[4934]: I1002 09:50:45.987839 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.015750 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.026985 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.046987 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.066634 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.086895 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.107368 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.127169 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.147414 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.167802 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.187817 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.207626 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.226723 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.248256 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.267286 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.286383 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.306876 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.326406 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.346135 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.367062 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.385070 4934 request.go:700] Waited for 1.004281386s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dpprof-cert&limit=500&resourceVersion=0 Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.387505 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.407611 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.426721 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.447163 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.466991 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.486892 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.506483 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.527043 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.546272 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.566767 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.587168 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.607489 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.626927 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.647857 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.666845 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.686566 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.707023 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.727022 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.747510 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.767487 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.786538 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.807181 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.826966 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.847374 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.866944 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.893497 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.907125 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.913930 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.927482 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.946883 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.966435 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 02 09:50:46 crc kubenswrapper[4934]: I1002 09:50:46.987091 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.007070 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.027225 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.046705 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.066718 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.087770 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.130603 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9j26x\" (UniqueName: \"kubernetes.io/projected/62e9cfe8-c134-423b-b971-c552f98526f9-kube-api-access-9j26x\") pod \"cluster-samples-operator-665b6dd947-gmxzf\" (UID: \"62e9cfe8-c134-423b-b971-c552f98526f9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gmxzf" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.146452 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4d8hf\" (UniqueName: \"kubernetes.io/projected/1d074a40-447d-4144-a2dc-ca69269c84de-kube-api-access-4d8hf\") pod \"apiserver-76f77b778f-fc4l5\" (UID: \"1d074a40-447d-4144-a2dc-ca69269c84de\") " pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.163057 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svdvd\" (UniqueName: \"kubernetes.io/projected/87b8c00f-e910-42f0-8089-73df46c6cf0b-kube-api-access-svdvd\") pod \"route-controller-manager-6576b87f9c-d6cv6\" (UID: \"87b8c00f-e910-42f0-8089-73df46c6cf0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.164755 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.182662 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn9w7\" (UniqueName: \"kubernetes.io/projected/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-kube-api-access-zn9w7\") pod \"controller-manager-879f6c89f-q2p9l\" (UID: \"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q2p9l" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.184011 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-q2p9l" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.189906 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gmxzf" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.207607 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.210047 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rchq\" (UniqueName: \"kubernetes.io/projected/8bbc993a-a1ea-48ae-a765-8f656d17329c-kube-api-access-9rchq\") pod \"machine-api-operator-5694c8668f-t6dmx\" (UID: \"8bbc993a-a1ea-48ae-a765-8f656d17329c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-t6dmx" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.224478 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hz4wv\" (UniqueName: \"kubernetes.io/projected/6096bf5b-7d9b-41c1-a6b9-d2f6879175a6-kube-api-access-hz4wv\") pod \"apiserver-7bbb656c7d-lfhf5\" (UID: \"6096bf5b-7d9b-41c1-a6b9-d2f6879175a6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.264299 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fwxz\" (UniqueName: \"kubernetes.io/projected/76e57d57-21b3-4f79-a7ee-b6bf6117c8d3-kube-api-access-2fwxz\") pod \"openshift-config-operator-7777fb866f-vhxn6\" (UID: \"76e57d57-21b3-4f79-a7ee-b6bf6117c8d3\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vhxn6" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.280850 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-t6dmx" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.283961 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jplkg\" (UniqueName: \"kubernetes.io/projected/602dc2aa-bffc-4e88-95ad-5dce0fc6d6df-kube-api-access-jplkg\") pod \"machine-approver-56656f9798-4bqtk\" (UID: \"602dc2aa-bffc-4e88-95ad-5dce0fc6d6df\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4bqtk" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.286692 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.309519 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.317925 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vhxn6" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.326888 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.348821 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.367978 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.369432 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4bqtk" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.372279 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gmxzf"] Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.387018 4934 request.go:700] Waited for 1.865781728s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns/secrets?fieldSelector=metadata.name%3Ddns-default-metrics-tls&limit=500&resourceVersion=0 Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.389289 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.392481 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.408009 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.413638 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-fc4l5"] Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.430910 4934 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.447246 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.452528 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q2p9l"] Oct 02 09:50:47 crc kubenswrapper[4934]: W1002 09:50:47.465256 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9e241d8_3d16_4cce_bfec_46d4b3bcb7b9.slice/crio-85bc11ec45ad9c47972b55d4674346fd4e8da517758ae63b529627b13151b7fa WatchSource:0}: Error finding container 85bc11ec45ad9c47972b55d4674346fd4e8da517758ae63b529627b13151b7fa: Status 404 returned error can't find the container with id 85bc11ec45ad9c47972b55d4674346fd4e8da517758ae63b529627b13151b7fa Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.469651 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.488660 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.491235 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-t6dmx"] Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.506408 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4bqtk" event={"ID":"602dc2aa-bffc-4e88-95ad-5dce0fc6d6df","Type":"ContainerStarted","Data":"6f165e061f04efd25456d3485d14c599b987a339d11c085500df0c52b2635928"} Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.509000 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-q2p9l" event={"ID":"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9","Type":"ContainerStarted","Data":"85bc11ec45ad9c47972b55d4674346fd4e8da517758ae63b529627b13151b7fa"} Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.509940 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gmxzf" event={"ID":"62e9cfe8-c134-423b-b971-c552f98526f9","Type":"ContainerStarted","Data":"0faa39bacbdc7fc8554a57b040bd1a72997a063a7fec37c6dbb3ae9eb10cad15"} Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.512368 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.512770 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" event={"ID":"1d074a40-447d-4144-a2dc-ca69269c84de","Type":"ContainerStarted","Data":"a6164172744733d62aab103bad05167638b74e1d0c127e5eed1ff9c00990aa5c"} Oct 02 09:50:47 crc kubenswrapper[4934]: W1002 09:50:47.521555 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8bbc993a_a1ea_48ae_a765_8f656d17329c.slice/crio-e30dd351f4f91d5e6f7c41828bdcad1259414dd0ec1554032a7d5d82bcab73dd WatchSource:0}: Error finding container e30dd351f4f91d5e6f7c41828bdcad1259414dd0ec1554032a7d5d82bcab73dd: Status 404 returned error can't find the container with id e30dd351f4f91d5e6f7c41828bdcad1259414dd0ec1554032a7d5d82bcab73dd Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.526112 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.541420 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-vhxn6"] Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.568176 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.586496 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.613806 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5"] Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.626976 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6"] Oct 02 09:50:47 crc kubenswrapper[4934]: W1002 09:50:47.627532 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6096bf5b_7d9b_41c1_a6b9_d2f6879175a6.slice/crio-b025eea4bad32b1be23d5b905a056c7f560134ae95b968aca51d37d9e95d4ff8 WatchSource:0}: Error finding container b025eea4bad32b1be23d5b905a056c7f560134ae95b968aca51d37d9e95d4ff8: Status 404 returned error can't find the container with id b025eea4bad32b1be23d5b905a056c7f560134ae95b968aca51d37d9e95d4ff8 Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.629479 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdxxv\" (UniqueName: \"kubernetes.io/projected/aa8f023f-df93-4883-9771-b0b23c472246-kube-api-access-rdxxv\") pod \"kube-storage-version-migrator-operator-b67b599dd-hb67r\" (UID: \"aa8f023f-df93-4883-9771-b0b23c472246\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb67r" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.629623 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/39e0228e-f4f6-4167-9bbc-d319374345a1-ca-trust-extracted\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.629707 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad155aa0-50a3-4874-93bd-2166a8d093cb-trusted-ca-bundle\") pod \"console-f9d7485db-f44r4\" (UID: \"ad155aa0-50a3-4874-93bd-2166a8d093cb\") " pod="openshift-console/console-f9d7485db-f44r4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.629779 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ad155aa0-50a3-4874-93bd-2166a8d093cb-service-ca\") pod \"console-f9d7485db-f44r4\" (UID: \"ad155aa0-50a3-4874-93bd-2166a8d093cb\") " pod="openshift-console/console-f9d7485db-f44r4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.629902 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.629993 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da6bd676-392c-4a8a-8a05-dee3304ef594-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-pxlv9\" (UID: \"da6bd676-392c-4a8a-8a05-dee3304ef594\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxlv9" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.630085 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ad155aa0-50a3-4874-93bd-2166a8d093cb-oauth-serving-cert\") pod \"console-f9d7485db-f44r4\" (UID: \"ad155aa0-50a3-4874-93bd-2166a8d093cb\") " pod="openshift-console/console-f9d7485db-f44r4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.630298 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-audit-policies\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.630361 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.630544 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fbf54dc7-2324-4259-b987-a0df48e06da5-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7qgvq\" (UID: \"fbf54dc7-2324-4259-b987-a0df48e06da5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7qgvq" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.630657 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/39e0228e-f4f6-4167-9bbc-d319374345a1-registry-certificates\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.630687 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d88b\" (UniqueName: \"kubernetes.io/projected/5dbb4e0e-9898-44ab-81f9-d149b80cfef0-kube-api-access-8d88b\") pod \"downloads-7954f5f757-mcnd2\" (UID: \"5dbb4e0e-9898-44ab-81f9-d149b80cfef0\") " pod="openshift-console/downloads-7954f5f757-mcnd2" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.630705 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ad155aa0-50a3-4874-93bd-2166a8d093cb-console-serving-cert\") pod \"console-f9d7485db-f44r4\" (UID: \"ad155aa0-50a3-4874-93bd-2166a8d093cb\") " pod="openshift-console/console-f9d7485db-f44r4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.630721 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8391ee1a-9c57-4745-825f-381f47254807-default-certificate\") pod \"router-default-5444994796-wjz5g\" (UID: \"8391ee1a-9c57-4745-825f-381f47254807\") " pod="openshift-ingress/router-default-5444994796-wjz5g" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.630740 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrbrd\" (UniqueName: \"kubernetes.io/projected/fec65034-065a-466e-8c61-dacc1e415b6f-kube-api-access-vrbrd\") pod \"openshift-controller-manager-operator-756b6f6bc6-7t8rp\" (UID: \"fec65034-065a-466e-8c61-dacc1e415b6f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7t8rp" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.630759 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa8f023f-df93-4883-9771-b0b23c472246-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-hb67r\" (UID: \"aa8f023f-df93-4883-9771-b0b23c472246\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb67r" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.630802 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc58m\" (UniqueName: \"kubernetes.io/projected/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-kube-api-access-rc58m\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.630833 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ebd4ee43-4dc5-46dc-964b-236cd1b16ec4-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-qg98s\" (UID: \"ebd4ee43-4dc5-46dc-964b-236cd1b16ec4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qg98s" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.630849 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-audit-dir\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.630865 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8391ee1a-9c57-4745-825f-381f47254807-metrics-certs\") pod \"router-default-5444994796-wjz5g\" (UID: \"8391ee1a-9c57-4745-825f-381f47254807\") " pod="openshift-ingress/router-default-5444994796-wjz5g" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.630881 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55tcl\" (UniqueName: \"kubernetes.io/projected/3978218d-91e2-4b20-8bcb-d1d77f820d88-kube-api-access-55tcl\") pod \"cluster-image-registry-operator-dc59b4c8b-6zlqv\" (UID: \"3978218d-91e2-4b20-8bcb-d1d77f820d88\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6zlqv" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.630906 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8391ee1a-9c57-4745-825f-381f47254807-service-ca-bundle\") pod \"router-default-5444994796-wjz5g\" (UID: \"8391ee1a-9c57-4745-825f-381f47254807\") " pod="openshift-ingress/router-default-5444994796-wjz5g" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.630923 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5f7c01ef-b9b0-476b-b3e0-12b8132b58b9-trusted-ca\") pod \"console-operator-58897d9998-5wrws\" (UID: \"5f7c01ef-b9b0-476b-b3e0-12b8132b58b9\") " pod="openshift-console-operator/console-operator-58897d9998-5wrws" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.631044 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ebd4ee43-4dc5-46dc-964b-236cd1b16ec4-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-qg98s\" (UID: \"ebd4ee43-4dc5-46dc-964b-236cd1b16ec4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qg98s" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.631105 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ad155aa0-50a3-4874-93bd-2166a8d093cb-console-config\") pod \"console-f9d7485db-f44r4\" (UID: \"ad155aa0-50a3-4874-93bd-2166a8d093cb\") " pod="openshift-console/console-f9d7485db-f44r4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.631175 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kr89h\" (UniqueName: \"kubernetes.io/projected/ad155aa0-50a3-4874-93bd-2166a8d093cb-kube-api-access-kr89h\") pod \"console-f9d7485db-f44r4\" (UID: \"ad155aa0-50a3-4874-93bd-2166a8d093cb\") " pod="openshift-console/console-f9d7485db-f44r4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.631546 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.631692 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da6bd676-392c-4a8a-8a05-dee3304ef594-config\") pod \"openshift-apiserver-operator-796bbdcf4f-pxlv9\" (UID: \"da6bd676-392c-4a8a-8a05-dee3304ef594\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxlv9" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.631849 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t857q\" (UniqueName: \"kubernetes.io/projected/460de405-cddd-4e59-8be4-537419c25f55-kube-api-access-t857q\") pod \"dns-operator-744455d44c-rjhkx\" (UID: \"460de405-cddd-4e59-8be4-537419c25f55\") " pod="openshift-dns-operator/dns-operator-744455d44c-rjhkx" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.631936 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.632066 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.632145 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.632261 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/39e0228e-f4f6-4167-9bbc-d319374345a1-bound-sa-token\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.632349 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/39e0228e-f4f6-4167-9bbc-d319374345a1-trusted-ca\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.632438 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66c8v\" (UniqueName: \"kubernetes.io/projected/da6bd676-392c-4a8a-8a05-dee3304ef594-kube-api-access-66c8v\") pod \"openshift-apiserver-operator-796bbdcf4f-pxlv9\" (UID: \"da6bd676-392c-4a8a-8a05-dee3304ef594\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxlv9" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.632505 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fec65034-065a-466e-8c61-dacc1e415b6f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-7t8rp\" (UID: \"fec65034-065a-466e-8c61-dacc1e415b6f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7t8rp" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.632631 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/39e0228e-f4f6-4167-9bbc-d319374345a1-registry-tls\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:47 crc kubenswrapper[4934]: E1002 09:50:47.632642 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:48.13262667 +0000 UTC m=+119.885268192 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.632866 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/39e0228e-f4f6-4167-9bbc-d319374345a1-installation-pull-secrets\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.632945 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f7c01ef-b9b0-476b-b3e0-12b8132b58b9-serving-cert\") pod \"console-operator-58897d9998-5wrws\" (UID: \"5f7c01ef-b9b0-476b-b3e0-12b8132b58b9\") " pod="openshift-console-operator/console-operator-58897d9998-5wrws" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.633055 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.633156 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbf54dc7-2324-4259-b987-a0df48e06da5-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7qgvq\" (UID: \"fbf54dc7-2324-4259-b987-a0df48e06da5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7qgvq" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.633233 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/586fee81-a26a-4506-8dfa-66db12798618-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-bnvc4\" (UID: \"586fee81-a26a-4506-8dfa-66db12798618\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bnvc4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.633301 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6d137db5-66a8-4066-b8d2-e88adcaaf12c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-66hrr\" (UID: \"6d137db5-66a8-4066-b8d2-e88adcaaf12c\") " pod="openshift-marketplace/marketplace-operator-79b997595-66hrr" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.633407 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3978218d-91e2-4b20-8bcb-d1d77f820d88-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-6zlqv\" (UID: \"3978218d-91e2-4b20-8bcb-d1d77f820d88\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6zlqv" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.633816 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l42zk\" (UniqueName: \"kubernetes.io/projected/6d137db5-66a8-4066-b8d2-e88adcaaf12c-kube-api-access-l42zk\") pod \"marketplace-operator-79b997595-66hrr\" (UID: \"6d137db5-66a8-4066-b8d2-e88adcaaf12c\") " pod="openshift-marketplace/marketplace-operator-79b997595-66hrr" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.633948 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-947cc\" (UniqueName: \"kubernetes.io/projected/8391ee1a-9c57-4745-825f-381f47254807-kube-api-access-947cc\") pod \"router-default-5444994796-wjz5g\" (UID: \"8391ee1a-9c57-4745-825f-381f47254807\") " pod="openshift-ingress/router-default-5444994796-wjz5g" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.634042 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/586fee81-a26a-4506-8dfa-66db12798618-config\") pod \"kube-apiserver-operator-766d6c64bb-bnvc4\" (UID: \"586fee81-a26a-4506-8dfa-66db12798618\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bnvc4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.634112 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/586fee81-a26a-4506-8dfa-66db12798618-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-bnvc4\" (UID: \"586fee81-a26a-4506-8dfa-66db12798618\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bnvc4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.634236 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.634307 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.634425 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aa8f023f-df93-4883-9771-b0b23c472246-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-hb67r\" (UID: \"aa8f023f-df93-4883-9771-b0b23c472246\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb67r" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.634540 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8391ee1a-9c57-4745-825f-381f47254807-stats-auth\") pod \"router-default-5444994796-wjz5g\" (UID: \"8391ee1a-9c57-4745-825f-381f47254807\") " pod="openshift-ingress/router-default-5444994796-wjz5g" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.634689 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/460de405-cddd-4e59-8be4-537419c25f55-metrics-tls\") pod \"dns-operator-744455d44c-rjhkx\" (UID: \"460de405-cddd-4e59-8be4-537419c25f55\") " pod="openshift-dns-operator/dns-operator-744455d44c-rjhkx" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.634819 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebd4ee43-4dc5-46dc-964b-236cd1b16ec4-config\") pod \"kube-controller-manager-operator-78b949d7b-qg98s\" (UID: \"ebd4ee43-4dc5-46dc-964b-236cd1b16ec4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qg98s" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.634944 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ad155aa0-50a3-4874-93bd-2166a8d093cb-console-oauth-config\") pod \"console-f9d7485db-f44r4\" (UID: \"ad155aa0-50a3-4874-93bd-2166a8d093cb\") " pod="openshift-console/console-f9d7485db-f44r4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.635062 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3978218d-91e2-4b20-8bcb-d1d77f820d88-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-6zlqv\" (UID: \"3978218d-91e2-4b20-8bcb-d1d77f820d88\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6zlqv" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.635160 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.635276 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.635848 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6d137db5-66a8-4066-b8d2-e88adcaaf12c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-66hrr\" (UID: \"6d137db5-66a8-4066-b8d2-e88adcaaf12c\") " pod="openshift-marketplace/marketplace-operator-79b997595-66hrr" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.635918 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f7c01ef-b9b0-476b-b3e0-12b8132b58b9-config\") pod \"console-operator-58897d9998-5wrws\" (UID: \"5f7c01ef-b9b0-476b-b3e0-12b8132b58b9\") " pod="openshift-console-operator/console-operator-58897d9998-5wrws" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.635950 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/3978218d-91e2-4b20-8bcb-d1d77f820d88-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-6zlqv\" (UID: \"3978218d-91e2-4b20-8bcb-d1d77f820d88\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6zlqv" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.635994 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.636029 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fec65034-065a-466e-8c61-dacc1e415b6f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-7t8rp\" (UID: \"fec65034-065a-466e-8c61-dacc1e415b6f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7t8rp" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.636074 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvmvq\" (UniqueName: \"kubernetes.io/projected/39e0228e-f4f6-4167-9bbc-d319374345a1-kube-api-access-fvmvq\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.636101 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b47xn\" (UniqueName: \"kubernetes.io/projected/5f7c01ef-b9b0-476b-b3e0-12b8132b58b9-kube-api-access-b47xn\") pod \"console-operator-58897d9998-5wrws\" (UID: \"5f7c01ef-b9b0-476b-b3e0-12b8132b58b9\") " pod="openshift-console-operator/console-operator-58897d9998-5wrws" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.636135 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fbf54dc7-2324-4259-b987-a0df48e06da5-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7qgvq\" (UID: \"fbf54dc7-2324-4259-b987-a0df48e06da5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7qgvq" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737011 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737193 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ebd4ee43-4dc5-46dc-964b-236cd1b16ec4-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-qg98s\" (UID: \"ebd4ee43-4dc5-46dc-964b-236cd1b16ec4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qg98s" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737215 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/327b634c-c056-49ff-a0ad-a178bc67ac48-apiservice-cert\") pod \"packageserver-d55dfcdfc-wlf9l\" (UID: \"327b634c-c056-49ff-a0ad-a178bc67ac48\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wlf9l" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737232 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/f4baf828-a5bb-4d7e-a43d-1194d47e4051-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5s688\" (UID: \"f4baf828-a5bb-4d7e-a43d-1194d47e4051\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5s688" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737250 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5q62x\" (UniqueName: \"kubernetes.io/projected/4b619ff7-c8d0-43ea-9373-18c36b2e160e-kube-api-access-5q62x\") pod \"authentication-operator-69f744f599-llqjv\" (UID: \"4b619ff7-c8d0-43ea-9373-18c36b2e160e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-llqjv" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737267 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e12b431-672f-404d-99ba-de5f718f19f2-serving-cert\") pod \"etcd-operator-b45778765-86wbg\" (UID: \"9e12b431-672f-404d-99ba-de5f718f19f2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-86wbg" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737283 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8391ee1a-9c57-4745-825f-381f47254807-metrics-certs\") pod \"router-default-5444994796-wjz5g\" (UID: \"8391ee1a-9c57-4745-825f-381f47254807\") " pod="openshift-ingress/router-default-5444994796-wjz5g" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737301 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5f7c01ef-b9b0-476b-b3e0-12b8132b58b9-trusted-ca\") pod \"console-operator-58897d9998-5wrws\" (UID: \"5f7c01ef-b9b0-476b-b3e0-12b8132b58b9\") " pod="openshift-console-operator/console-operator-58897d9998-5wrws" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737316 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tb84s\" (UniqueName: \"kubernetes.io/projected/90d8cb96-af2e-41b8-a405-85cf9e017631-kube-api-access-tb84s\") pod \"multus-admission-controller-857f4d67dd-hr945\" (UID: \"90d8cb96-af2e-41b8-a405-85cf9e017631\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hr945" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737340 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ad155aa0-50a3-4874-93bd-2166a8d093cb-console-config\") pod \"console-f9d7485db-f44r4\" (UID: \"ad155aa0-50a3-4874-93bd-2166a8d093cb\") " pod="openshift-console/console-f9d7485db-f44r4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737355 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jfwf\" (UniqueName: \"kubernetes.io/projected/59279e86-a90f-463e-a098-da83a2bdf5a5-kube-api-access-7jfwf\") pod \"machine-config-controller-84d6567774-6jtqh\" (UID: \"59279e86-a90f-463e-a098-da83a2bdf5a5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6jtqh" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737370 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l46wc\" (UniqueName: \"kubernetes.io/projected/dcb231bb-d375-49d3-b967-dc739da957fb-kube-api-access-l46wc\") pod \"dns-default-v58f6\" (UID: \"dcb231bb-d375-49d3-b967-dc739da957fb\") " pod="openshift-dns/dns-default-v58f6" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737386 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737400 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e12b431-672f-404d-99ba-de5f718f19f2-config\") pod \"etcd-operator-b45778765-86wbg\" (UID: \"9e12b431-672f-404d-99ba-de5f718f19f2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-86wbg" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737416 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/9e12b431-672f-404d-99ba-de5f718f19f2-etcd-ca\") pod \"etcd-operator-b45778765-86wbg\" (UID: \"9e12b431-672f-404d-99ba-de5f718f19f2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-86wbg" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737432 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/327b634c-c056-49ff-a0ad-a178bc67ac48-tmpfs\") pod \"packageserver-d55dfcdfc-wlf9l\" (UID: \"327b634c-c056-49ff-a0ad-a178bc67ac48\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wlf9l" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737448 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/de99fcca-2d10-4595-8827-410ac05c5f68-registration-dir\") pod \"csi-hostpathplugin-pfl7n\" (UID: \"de99fcca-2d10-4595-8827-410ac05c5f68\") " pod="hostpath-provisioner/csi-hostpathplugin-pfl7n" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737475 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/39e0228e-f4f6-4167-9bbc-d319374345a1-bound-sa-token\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737669 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dcb231bb-d375-49d3-b967-dc739da957fb-config-volume\") pod \"dns-default-v58f6\" (UID: \"dcb231bb-d375-49d3-b967-dc739da957fb\") " pod="openshift-dns/dns-default-v58f6" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737697 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fec65034-065a-466e-8c61-dacc1e415b6f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-7t8rp\" (UID: \"fec65034-065a-466e-8c61-dacc1e415b6f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7t8rp" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737719 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b52j8\" (UniqueName: \"kubernetes.io/projected/ef34af0b-654c-4519-9292-4e1a833b2ed3-kube-api-access-b52j8\") pod \"control-plane-machine-set-operator-78cbb6b69f-z2d6w\" (UID: \"ef34af0b-654c-4519-9292-4e1a833b2ed3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z2d6w" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737745 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrp8h\" (UniqueName: \"kubernetes.io/projected/66183279-91fd-423e-91bd-19b68dda8ef0-kube-api-access-lrp8h\") pod \"catalog-operator-68c6474976-9ts9g\" (UID: \"66183279-91fd-423e-91bd-19b68dda8ef0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9ts9g" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737764 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/25f4d147-18a5-4d67-99c5-39273eecacc0-cert\") pod \"ingress-canary-pzrnw\" (UID: \"25f4d147-18a5-4d67-99c5-39273eecacc0\") " pod="openshift-ingress-canary/ingress-canary-pzrnw" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737784 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737804 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/39e0228e-f4f6-4167-9bbc-d319374345a1-registry-tls\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737834 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbf54dc7-2324-4259-b987-a0df48e06da5-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7qgvq\" (UID: \"fbf54dc7-2324-4259-b987-a0df48e06da5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7qgvq" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737864 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfzlb\" (UniqueName: \"kubernetes.io/projected/8549da74-9ee6-46ff-a0dc-5b4af7d6cfa6-kube-api-access-lfzlb\") pod \"service-ca-operator-777779d784-hw97b\" (UID: \"8549da74-9ee6-46ff-a0dc-5b4af7d6cfa6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hw97b" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737883 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/9e12b431-672f-404d-99ba-de5f718f19f2-etcd-service-ca\") pod \"etcd-operator-b45778765-86wbg\" (UID: \"9e12b431-672f-404d-99ba-de5f718f19f2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-86wbg" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737900 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/66183279-91fd-423e-91bd-19b68dda8ef0-srv-cert\") pod \"catalog-operator-68c6474976-9ts9g\" (UID: \"66183279-91fd-423e-91bd-19b68dda8ef0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9ts9g" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737953 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3978218d-91e2-4b20-8bcb-d1d77f820d88-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-6zlqv\" (UID: \"3978218d-91e2-4b20-8bcb-d1d77f820d88\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6zlqv" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.737980 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/90d8cb96-af2e-41b8-a405-85cf9e017631-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-hr945\" (UID: \"90d8cb96-af2e-41b8-a405-85cf9e017631\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hr945" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.738001 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pshs9\" (UniqueName: \"kubernetes.io/projected/ad1a290c-1e04-46fe-8118-92bab2478385-kube-api-access-pshs9\") pod \"ingress-operator-5b745b69d9-x7mrm\" (UID: \"ad1a290c-1e04-46fe-8118-92bab2478385\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-x7mrm" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.738025 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-947cc\" (UniqueName: \"kubernetes.io/projected/8391ee1a-9c57-4745-825f-381f47254807-kube-api-access-947cc\") pod \"router-default-5444994796-wjz5g\" (UID: \"8391ee1a-9c57-4745-825f-381f47254807\") " pod="openshift-ingress/router-default-5444994796-wjz5g" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.738046 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/586fee81-a26a-4506-8dfa-66db12798618-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-bnvc4\" (UID: \"586fee81-a26a-4506-8dfa-66db12798618\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bnvc4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.738066 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/88ecaab6-9c13-4813-8698-5b558d32a2e9-srv-cert\") pod \"olm-operator-6b444d44fb-lch8r\" (UID: \"88ecaab6-9c13-4813-8698-5b558d32a2e9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lch8r" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.738101 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aa8f023f-df93-4883-9771-b0b23c472246-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-hb67r\" (UID: \"aa8f023f-df93-4883-9771-b0b23c472246\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb67r" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.738122 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a-secret-volume\") pod \"collect-profiles-29323305-jv64v\" (UID: \"c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-jv64v" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.738156 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrhkt\" (UniqueName: \"kubernetes.io/projected/25f4d147-18a5-4d67-99c5-39273eecacc0-kube-api-access-vrhkt\") pod \"ingress-canary-pzrnw\" (UID: \"25f4d147-18a5-4d67-99c5-39273eecacc0\") " pod="openshift-ingress-canary/ingress-canary-pzrnw" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.738180 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsgsr\" (UniqueName: \"kubernetes.io/projected/c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a-kube-api-access-vsgsr\") pod \"collect-profiles-29323305-jv64v\" (UID: \"c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-jv64v" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.738214 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3978218d-91e2-4b20-8bcb-d1d77f820d88-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-6zlqv\" (UID: \"3978218d-91e2-4b20-8bcb-d1d77f820d88\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6zlqv" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.738237 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.738258 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/88ecaab6-9c13-4813-8698-5b558d32a2e9-profile-collector-cert\") pod \"olm-operator-6b444d44fb-lch8r\" (UID: \"88ecaab6-9c13-4813-8698-5b558d32a2e9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lch8r" Oct 02 09:50:47 crc kubenswrapper[4934]: E1002 09:50:47.738275 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:48.238257088 +0000 UTC m=+119.990898610 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.738294 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f7c01ef-b9b0-476b-b3e0-12b8132b58b9-config\") pod \"console-operator-58897d9998-5wrws\" (UID: \"5f7c01ef-b9b0-476b-b3e0-12b8132b58b9\") " pod="openshift-console-operator/console-operator-58897d9998-5wrws" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.738315 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/9fd0eab0-2348-495b-83cf-580fedb4c1f1-certs\") pod \"machine-config-server-w6z57\" (UID: \"9fd0eab0-2348-495b-83cf-580fedb4c1f1\") " pod="openshift-machine-config-operator/machine-config-server-w6z57" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.738333 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fec65034-065a-466e-8c61-dacc1e415b6f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-7t8rp\" (UID: \"fec65034-065a-466e-8c61-dacc1e415b6f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7t8rp" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.741493 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fec65034-065a-466e-8c61-dacc1e415b6f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-7t8rp\" (UID: \"fec65034-065a-466e-8c61-dacc1e415b6f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7t8rp" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.741795 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7kk9\" (UniqueName: \"kubernetes.io/projected/28aceb4d-e776-4340-9812-8150b2766d5c-kube-api-access-r7kk9\") pod \"migrator-59844c95c7-8xjjw\" (UID: \"28aceb4d-e776-4340-9812-8150b2766d5c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8xjjw" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.741847 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/327b634c-c056-49ff-a0ad-a178bc67ac48-webhook-cert\") pod \"packageserver-d55dfcdfc-wlf9l\" (UID: \"327b634c-c056-49ff-a0ad-a178bc67ac48\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wlf9l" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.741920 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.741942 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/59279e86-a90f-463e-a098-da83a2bdf5a5-proxy-tls\") pod \"machine-config-controller-84d6567774-6jtqh\" (UID: \"59279e86-a90f-463e-a098-da83a2bdf5a5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6jtqh" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.741967 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f7c01ef-b9b0-476b-b3e0-12b8132b58b9-config\") pod \"console-operator-58897d9998-5wrws\" (UID: \"5f7c01ef-b9b0-476b-b3e0-12b8132b58b9\") " pod="openshift-console-operator/console-operator-58897d9998-5wrws" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.741975 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b619ff7-c8d0-43ea-9373-18c36b2e160e-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-llqjv\" (UID: \"4b619ff7-c8d0-43ea-9373-18c36b2e160e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-llqjv" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.741994 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ad1a290c-1e04-46fe-8118-92bab2478385-bound-sa-token\") pod \"ingress-operator-5b745b69d9-x7mrm\" (UID: \"ad1a290c-1e04-46fe-8118-92bab2478385\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-x7mrm" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.742018 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdxxv\" (UniqueName: \"kubernetes.io/projected/aa8f023f-df93-4883-9771-b0b23c472246-kube-api-access-rdxxv\") pod \"kube-storage-version-migrator-operator-b67b599dd-hb67r\" (UID: \"aa8f023f-df93-4883-9771-b0b23c472246\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb67r" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.742041 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/305740e5-7461-4097-994e-731594c096bb-auth-proxy-config\") pod \"machine-config-operator-74547568cd-7frl9\" (UID: \"305740e5-7461-4097-994e-731594c096bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7frl9" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.743001 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.743053 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbf54dc7-2324-4259-b987-a0df48e06da5-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7qgvq\" (UID: \"fbf54dc7-2324-4259-b987-a0df48e06da5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7qgvq" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.743344 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5f7c01ef-b9b0-476b-b3e0-12b8132b58b9-trusted-ca\") pod \"console-operator-58897d9998-5wrws\" (UID: \"5f7c01ef-b9b0-476b-b3e0-12b8132b58b9\") " pod="openshift-console-operator/console-operator-58897d9998-5wrws" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.744028 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/39e0228e-f4f6-4167-9bbc-d319374345a1-ca-trust-extracted\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.744087 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a-config-volume\") pod \"collect-profiles-29323305-jv64v\" (UID: \"c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-jv64v" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.744126 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8549da74-9ee6-46ff-a0dc-5b4af7d6cfa6-config\") pod \"service-ca-operator-777779d784-hw97b\" (UID: \"8549da74-9ee6-46ff-a0dc-5b4af7d6cfa6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hw97b" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.744774 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/39e0228e-f4f6-4167-9bbc-d319374345a1-ca-trust-extracted\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.744959 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ad155aa0-50a3-4874-93bd-2166a8d093cb-console-config\") pod \"console-f9d7485db-f44r4\" (UID: \"ad155aa0-50a3-4874-93bd-2166a8d093cb\") " pod="openshift-console/console-f9d7485db-f44r4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.745655 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad155aa0-50a3-4874-93bd-2166a8d093cb-trusted-ca-bundle\") pod \"console-f9d7485db-f44r4\" (UID: \"ad155aa0-50a3-4874-93bd-2166a8d093cb\") " pod="openshift-console/console-f9d7485db-f44r4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.746618 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.746767 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3978218d-91e2-4b20-8bcb-d1d77f820d88-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-6zlqv\" (UID: \"3978218d-91e2-4b20-8bcb-d1d77f820d88\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6zlqv" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.747049 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad155aa0-50a3-4874-93bd-2166a8d093cb-trusted-ca-bundle\") pod \"console-f9d7485db-f44r4\" (UID: \"ad155aa0-50a3-4874-93bd-2166a8d093cb\") " pod="openshift-console/console-f9d7485db-f44r4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.750675 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aa8f023f-df93-4883-9771-b0b23c472246-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-hb67r\" (UID: \"aa8f023f-df93-4883-9771-b0b23c472246\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb67r" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.752291 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.752561 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxcmq\" (UniqueName: \"kubernetes.io/projected/8f6c0ab1-dd58-4734-89f8-aefb606078d3-kube-api-access-pxcmq\") pod \"service-ca-9c57cc56f-2vn7r\" (UID: \"8f6c0ab1-dd58-4734-89f8-aefb606078d3\") " pod="openshift-service-ca/service-ca-9c57cc56f-2vn7r" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.752629 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.752668 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-audit-policies\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.752687 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.752701 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/39e0228e-f4f6-4167-9bbc-d319374345a1-registry-certificates\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.752796 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d88b\" (UniqueName: \"kubernetes.io/projected/5dbb4e0e-9898-44ab-81f9-d149b80cfef0-kube-api-access-8d88b\") pod \"downloads-7954f5f757-mcnd2\" (UID: \"5dbb4e0e-9898-44ab-81f9-d149b80cfef0\") " pod="openshift-console/downloads-7954f5f757-mcnd2" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.752822 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ad155aa0-50a3-4874-93bd-2166a8d093cb-console-serving-cert\") pod \"console-f9d7485db-f44r4\" (UID: \"ad155aa0-50a3-4874-93bd-2166a8d093cb\") " pod="openshift-console/console-f9d7485db-f44r4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.752840 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ad155aa0-50a3-4874-93bd-2166a8d093cb-oauth-serving-cert\") pod \"console-f9d7485db-f44r4\" (UID: \"ad155aa0-50a3-4874-93bd-2166a8d093cb\") " pod="openshift-console/console-f9d7485db-f44r4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.752867 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa8f023f-df93-4883-9771-b0b23c472246-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-hb67r\" (UID: \"aa8f023f-df93-4883-9771-b0b23c472246\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb67r" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.752895 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/66183279-91fd-423e-91bd-19b68dda8ef0-profile-collector-cert\") pod \"catalog-operator-68c6474976-9ts9g\" (UID: \"66183279-91fd-423e-91bd-19b68dda8ef0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9ts9g" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.752925 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b619ff7-c8d0-43ea-9373-18c36b2e160e-config\") pod \"authentication-operator-69f744f599-llqjv\" (UID: \"4b619ff7-c8d0-43ea-9373-18c36b2e160e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-llqjv" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.752967 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-audit-dir\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.753008 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8391ee1a-9c57-4745-825f-381f47254807-service-ca-bundle\") pod \"router-default-5444994796-wjz5g\" (UID: \"8391ee1a-9c57-4745-825f-381f47254807\") " pod="openshift-ingress/router-default-5444994796-wjz5g" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.753028 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55tcl\" (UniqueName: \"kubernetes.io/projected/3978218d-91e2-4b20-8bcb-d1d77f820d88-kube-api-access-55tcl\") pod \"cluster-image-registry-operator-dc59b4c8b-6zlqv\" (UID: \"3978218d-91e2-4b20-8bcb-d1d77f820d88\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6zlqv" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.753056 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bd4h\" (UniqueName: \"kubernetes.io/projected/88ecaab6-9c13-4813-8698-5b558d32a2e9-kube-api-access-4bd4h\") pod \"olm-operator-6b444d44fb-lch8r\" (UID: \"88ecaab6-9c13-4813-8698-5b558d32a2e9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lch8r" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.753105 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ebd4ee43-4dc5-46dc-964b-236cd1b16ec4-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-qg98s\" (UID: \"ebd4ee43-4dc5-46dc-964b-236cd1b16ec4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qg98s" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.753129 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da6bd676-392c-4a8a-8a05-dee3304ef594-config\") pod \"openshift-apiserver-operator-796bbdcf4f-pxlv9\" (UID: \"da6bd676-392c-4a8a-8a05-dee3304ef594\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxlv9" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.753154 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/de99fcca-2d10-4595-8827-410ac05c5f68-mountpoint-dir\") pod \"csi-hostpathplugin-pfl7n\" (UID: \"de99fcca-2d10-4595-8827-410ac05c5f68\") " pod="hostpath-provisioner/csi-hostpathplugin-pfl7n" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.753175 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kr89h\" (UniqueName: \"kubernetes.io/projected/ad155aa0-50a3-4874-93bd-2166a8d093cb-kube-api-access-kr89h\") pod \"console-f9d7485db-f44r4\" (UID: \"ad155aa0-50a3-4874-93bd-2166a8d093cb\") " pod="openshift-console/console-f9d7485db-f44r4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.753198 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.753225 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.753250 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t857q\" (UniqueName: \"kubernetes.io/projected/460de405-cddd-4e59-8be4-537419c25f55-kube-api-access-t857q\") pod \"dns-operator-744455d44c-rjhkx\" (UID: \"460de405-cddd-4e59-8be4-537419c25f55\") " pod="openshift-dns-operator/dns-operator-744455d44c-rjhkx" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.753270 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/8f6c0ab1-dd58-4734-89f8-aefb606078d3-signing-cabundle\") pod \"service-ca-9c57cc56f-2vn7r\" (UID: \"8f6c0ab1-dd58-4734-89f8-aefb606078d3\") " pod="openshift-service-ca/service-ca-9c57cc56f-2vn7r" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.753305 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.753326 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/de99fcca-2d10-4595-8827-410ac05c5f68-socket-dir\") pod \"csi-hostpathplugin-pfl7n\" (UID: \"de99fcca-2d10-4595-8827-410ac05c5f68\") " pod="hostpath-provisioner/csi-hostpathplugin-pfl7n" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.753358 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/39e0228e-f4f6-4167-9bbc-d319374345a1-trusted-ca\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.753380 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrqkv\" (UniqueName: \"kubernetes.io/projected/f4baf828-a5bb-4d7e-a43d-1194d47e4051-kube-api-access-jrqkv\") pod \"package-server-manager-789f6589d5-5s688\" (UID: \"f4baf828-a5bb-4d7e-a43d-1194d47e4051\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5s688" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.753640 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66c8v\" (UniqueName: \"kubernetes.io/projected/da6bd676-392c-4a8a-8a05-dee3304ef594-kube-api-access-66c8v\") pod \"openshift-apiserver-operator-796bbdcf4f-pxlv9\" (UID: \"da6bd676-392c-4a8a-8a05-dee3304ef594\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxlv9" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.753687 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b619ff7-c8d0-43ea-9373-18c36b2e160e-serving-cert\") pod \"authentication-operator-69f744f599-llqjv\" (UID: \"4b619ff7-c8d0-43ea-9373-18c36b2e160e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-llqjv" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.753382 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fec65034-065a-466e-8c61-dacc1e415b6f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-7t8rp\" (UID: \"fec65034-065a-466e-8c61-dacc1e415b6f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7t8rp" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.753723 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/39e0228e-f4f6-4167-9bbc-d319374345a1-installation-pull-secrets\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.753925 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f7c01ef-b9b0-476b-b3e0-12b8132b58b9-serving-cert\") pod \"console-operator-58897d9998-5wrws\" (UID: \"5f7c01ef-b9b0-476b-b3e0-12b8132b58b9\") " pod="openshift-console-operator/console-operator-58897d9998-5wrws" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.753947 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/ef34af0b-654c-4519-9292-4e1a833b2ed3-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-z2d6w\" (UID: \"ef34af0b-654c-4519-9292-4e1a833b2ed3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z2d6w" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.753975 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbqk7\" (UniqueName: \"kubernetes.io/projected/9fd0eab0-2348-495b-83cf-580fedb4c1f1-kube-api-access-zbqk7\") pod \"machine-config-server-w6z57\" (UID: \"9fd0eab0-2348-495b-83cf-580fedb4c1f1\") " pod="openshift-machine-config-operator/machine-config-server-w6z57" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.754001 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/305740e5-7461-4097-994e-731594c096bb-proxy-tls\") pod \"machine-config-operator-74547568cd-7frl9\" (UID: \"305740e5-7461-4097-994e-731594c096bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7frl9" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.754069 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa8f023f-df93-4883-9771-b0b23c472246-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-hb67r\" (UID: \"aa8f023f-df93-4883-9771-b0b23c472246\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb67r" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.754255 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/586fee81-a26a-4506-8dfa-66db12798618-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-bnvc4\" (UID: \"586fee81-a26a-4506-8dfa-66db12798618\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bnvc4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.754303 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/39e0228e-f4f6-4167-9bbc-d319374345a1-registry-certificates\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.754280 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6d137db5-66a8-4066-b8d2-e88adcaaf12c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-66hrr\" (UID: \"6d137db5-66a8-4066-b8d2-e88adcaaf12c\") " pod="openshift-marketplace/marketplace-operator-79b997595-66hrr" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.754419 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/59279e86-a90f-463e-a098-da83a2bdf5a5-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-6jtqh\" (UID: \"59279e86-a90f-463e-a098-da83a2bdf5a5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6jtqh" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.754617 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jflw\" (UniqueName: \"kubernetes.io/projected/327b634c-c056-49ff-a0ad-a178bc67ac48-kube-api-access-6jflw\") pod \"packageserver-d55dfcdfc-wlf9l\" (UID: \"327b634c-c056-49ff-a0ad-a178bc67ac48\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wlf9l" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.754696 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ad1a290c-1e04-46fe-8118-92bab2478385-trusted-ca\") pod \"ingress-operator-5b745b69d9-x7mrm\" (UID: \"ad1a290c-1e04-46fe-8118-92bab2478385\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-x7mrm" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.754718 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l42zk\" (UniqueName: \"kubernetes.io/projected/6d137db5-66a8-4066-b8d2-e88adcaaf12c-kube-api-access-l42zk\") pod \"marketplace-operator-79b997595-66hrr\" (UID: \"6d137db5-66a8-4066-b8d2-e88adcaaf12c\") " pod="openshift-marketplace/marketplace-operator-79b997595-66hrr" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.754989 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ad1a290c-1e04-46fe-8118-92bab2478385-metrics-tls\") pod \"ingress-operator-5b745b69d9-x7mrm\" (UID: \"ad1a290c-1e04-46fe-8118-92bab2478385\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-x7mrm" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.755014 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/586fee81-a26a-4506-8dfa-66db12798618-config\") pod \"kube-apiserver-operator-766d6c64bb-bnvc4\" (UID: \"586fee81-a26a-4506-8dfa-66db12798618\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bnvc4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.755117 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-audit-policies\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.755245 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b619ff7-c8d0-43ea-9373-18c36b2e160e-service-ca-bundle\") pod \"authentication-operator-69f744f599-llqjv\" (UID: \"4b619ff7-c8d0-43ea-9373-18c36b2e160e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-llqjv" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.755270 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/9fd0eab0-2348-495b-83cf-580fedb4c1f1-node-bootstrap-token\") pod \"machine-config-server-w6z57\" (UID: \"9fd0eab0-2348-495b-83cf-580fedb4c1f1\") " pod="openshift-machine-config-operator/machine-config-server-w6z57" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.755583 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8391ee1a-9c57-4745-825f-381f47254807-stats-auth\") pod \"router-default-5444994796-wjz5g\" (UID: \"8391ee1a-9c57-4745-825f-381f47254807\") " pod="openshift-ingress/router-default-5444994796-wjz5g" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.755614 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.755770 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.755796 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/8f6c0ab1-dd58-4734-89f8-aefb606078d3-signing-key\") pod \"service-ca-9c57cc56f-2vn7r\" (UID: \"8f6c0ab1-dd58-4734-89f8-aefb606078d3\") " pod="openshift-service-ca/service-ca-9c57cc56f-2vn7r" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.755900 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/305740e5-7461-4097-994e-731594c096bb-images\") pod \"machine-config-operator-74547568cd-7frl9\" (UID: \"305740e5-7461-4097-994e-731594c096bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7frl9" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.755929 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ad155aa0-50a3-4874-93bd-2166a8d093cb-console-oauth-config\") pod \"console-f9d7485db-f44r4\" (UID: \"ad155aa0-50a3-4874-93bd-2166a8d093cb\") " pod="openshift-console/console-f9d7485db-f44r4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.755952 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/460de405-cddd-4e59-8be4-537419c25f55-metrics-tls\") pod \"dns-operator-744455d44c-rjhkx\" (UID: \"460de405-cddd-4e59-8be4-537419c25f55\") " pod="openshift-dns-operator/dns-operator-744455d44c-rjhkx" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.756150 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebd4ee43-4dc5-46dc-964b-236cd1b16ec4-config\") pod \"kube-controller-manager-operator-78b949d7b-qg98s\" (UID: \"ebd4ee43-4dc5-46dc-964b-236cd1b16ec4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qg98s" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.756174 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8549da74-9ee6-46ff-a0dc-5b4af7d6cfa6-serving-cert\") pod \"service-ca-operator-777779d784-hw97b\" (UID: \"8549da74-9ee6-46ff-a0dc-5b4af7d6cfa6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hw97b" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.756320 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/dcb231bb-d375-49d3-b967-dc739da957fb-metrics-tls\") pod \"dns-default-v58f6\" (UID: \"dcb231bb-d375-49d3-b967-dc739da957fb\") " pod="openshift-dns/dns-default-v58f6" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.756355 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/3978218d-91e2-4b20-8bcb-d1d77f820d88-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-6zlqv\" (UID: \"3978218d-91e2-4b20-8bcb-d1d77f820d88\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6zlqv" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.756235 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ad155aa0-50a3-4874-93bd-2166a8d093cb-oauth-serving-cert\") pod \"console-f9d7485db-f44r4\" (UID: \"ad155aa0-50a3-4874-93bd-2166a8d093cb\") " pod="openshift-console/console-f9d7485db-f44r4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.756384 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.756590 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6d137db5-66a8-4066-b8d2-e88adcaaf12c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-66hrr\" (UID: \"6d137db5-66a8-4066-b8d2-e88adcaaf12c\") " pod="openshift-marketplace/marketplace-operator-79b997595-66hrr" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.756809 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fbf54dc7-2324-4259-b987-a0df48e06da5-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7qgvq\" (UID: \"fbf54dc7-2324-4259-b987-a0df48e06da5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7qgvq" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.756885 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvmvq\" (UniqueName: \"kubernetes.io/projected/39e0228e-f4f6-4167-9bbc-d319374345a1-kube-api-access-fvmvq\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.757003 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.757047 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b47xn\" (UniqueName: \"kubernetes.io/projected/5f7c01ef-b9b0-476b-b3e0-12b8132b58b9-kube-api-access-b47xn\") pod \"console-operator-58897d9998-5wrws\" (UID: \"5f7c01ef-b9b0-476b-b3e0-12b8132b58b9\") " pod="openshift-console-operator/console-operator-58897d9998-5wrws" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.757075 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/de99fcca-2d10-4595-8827-410ac05c5f68-csi-data-dir\") pod \"csi-hostpathplugin-pfl7n\" (UID: \"de99fcca-2d10-4595-8827-410ac05c5f68\") " pod="hostpath-provisioner/csi-hostpathplugin-pfl7n" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.757224 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mll7c\" (UniqueName: \"kubernetes.io/projected/de99fcca-2d10-4595-8827-410ac05c5f68-kube-api-access-mll7c\") pod \"csi-hostpathplugin-pfl7n\" (UID: \"de99fcca-2d10-4595-8827-410ac05c5f68\") " pod="hostpath-provisioner/csi-hostpathplugin-pfl7n" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.757244 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/de99fcca-2d10-4595-8827-410ac05c5f68-plugins-dir\") pod \"csi-hostpathplugin-pfl7n\" (UID: \"de99fcca-2d10-4595-8827-410ac05c5f68\") " pod="hostpath-provisioner/csi-hostpathplugin-pfl7n" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.757674 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da6bd676-392c-4a8a-8a05-dee3304ef594-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-pxlv9\" (UID: \"da6bd676-392c-4a8a-8a05-dee3304ef594\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxlv9" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.757703 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9e12b431-672f-404d-99ba-de5f718f19f2-etcd-client\") pod \"etcd-operator-b45778765-86wbg\" (UID: \"9e12b431-672f-404d-99ba-de5f718f19f2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-86wbg" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.757871 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ad155aa0-50a3-4874-93bd-2166a8d093cb-service-ca\") pod \"console-f9d7485db-f44r4\" (UID: \"ad155aa0-50a3-4874-93bd-2166a8d093cb\") " pod="openshift-console/console-f9d7485db-f44r4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.757892 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.758050 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-audit-dir\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.757917 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fbf54dc7-2324-4259-b987-a0df48e06da5-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7qgvq\" (UID: \"fbf54dc7-2324-4259-b987-a0df48e06da5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7qgvq" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.758121 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmxrj\" (UniqueName: \"kubernetes.io/projected/9e12b431-672f-404d-99ba-de5f718f19f2-kube-api-access-zmxrj\") pod \"etcd-operator-b45778765-86wbg\" (UID: \"9e12b431-672f-404d-99ba-de5f718f19f2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-86wbg" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.758159 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bszcl\" (UniqueName: \"kubernetes.io/projected/305740e5-7461-4097-994e-731594c096bb-kube-api-access-bszcl\") pod \"machine-config-operator-74547568cd-7frl9\" (UID: \"305740e5-7461-4097-994e-731594c096bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7frl9" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.758526 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ad155aa0-50a3-4874-93bd-2166a8d093cb-console-serving-cert\") pod \"console-f9d7485db-f44r4\" (UID: \"ad155aa0-50a3-4874-93bd-2166a8d093cb\") " pod="openshift-console/console-f9d7485db-f44r4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.758660 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8391ee1a-9c57-4745-825f-381f47254807-metrics-certs\") pod \"router-default-5444994796-wjz5g\" (UID: \"8391ee1a-9c57-4745-825f-381f47254807\") " pod="openshift-ingress/router-default-5444994796-wjz5g" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.756317 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/39e0228e-f4f6-4167-9bbc-d319374345a1-registry-tls\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.759998 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8391ee1a-9c57-4745-825f-381f47254807-service-ca-bundle\") pod \"router-default-5444994796-wjz5g\" (UID: \"8391ee1a-9c57-4745-825f-381f47254807\") " pod="openshift-ingress/router-default-5444994796-wjz5g" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.761666 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/39e0228e-f4f6-4167-9bbc-d319374345a1-trusted-ca\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.762185 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ebd4ee43-4dc5-46dc-964b-236cd1b16ec4-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-qg98s\" (UID: \"ebd4ee43-4dc5-46dc-964b-236cd1b16ec4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qg98s" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.764216 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ad155aa0-50a3-4874-93bd-2166a8d093cb-service-ca\") pod \"console-f9d7485db-f44r4\" (UID: \"ad155aa0-50a3-4874-93bd-2166a8d093cb\") " pod="openshift-console/console-f9d7485db-f44r4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.764539 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.764584 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da6bd676-392c-4a8a-8a05-dee3304ef594-config\") pod \"openshift-apiserver-operator-796bbdcf4f-pxlv9\" (UID: \"da6bd676-392c-4a8a-8a05-dee3304ef594\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxlv9" Oct 02 09:50:47 crc kubenswrapper[4934]: E1002 09:50:47.765202 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:48.26517971 +0000 UTC m=+120.017821232 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.765511 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6d137db5-66a8-4066-b8d2-e88adcaaf12c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-66hrr\" (UID: \"6d137db5-66a8-4066-b8d2-e88adcaaf12c\") " pod="openshift-marketplace/marketplace-operator-79b997595-66hrr" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.765654 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8391ee1a-9c57-4745-825f-381f47254807-default-certificate\") pod \"router-default-5444994796-wjz5g\" (UID: \"8391ee1a-9c57-4745-825f-381f47254807\") " pod="openshift-ingress/router-default-5444994796-wjz5g" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.765683 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrbrd\" (UniqueName: \"kubernetes.io/projected/fec65034-065a-466e-8c61-dacc1e415b6f-kube-api-access-vrbrd\") pod \"openshift-controller-manager-operator-756b6f6bc6-7t8rp\" (UID: \"fec65034-065a-466e-8c61-dacc1e415b6f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7t8rp" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.765708 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc58m\" (UniqueName: \"kubernetes.io/projected/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-kube-api-access-rc58m\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.766419 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.767129 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/586fee81-a26a-4506-8dfa-66db12798618-config\") pod \"kube-apiserver-operator-766d6c64bb-bnvc4\" (UID: \"586fee81-a26a-4506-8dfa-66db12798618\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bnvc4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.767611 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.767631 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebd4ee43-4dc5-46dc-964b-236cd1b16ec4-config\") pod \"kube-controller-manager-operator-78b949d7b-qg98s\" (UID: \"ebd4ee43-4dc5-46dc-964b-236cd1b16ec4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qg98s" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.769093 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.772621 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/460de405-cddd-4e59-8be4-537419c25f55-metrics-tls\") pod \"dns-operator-744455d44c-rjhkx\" (UID: \"460de405-cddd-4e59-8be4-537419c25f55\") " pod="openshift-dns-operator/dns-operator-744455d44c-rjhkx" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.772724 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f7c01ef-b9b0-476b-b3e0-12b8132b58b9-serving-cert\") pod \"console-operator-58897d9998-5wrws\" (UID: \"5f7c01ef-b9b0-476b-b3e0-12b8132b58b9\") " pod="openshift-console-operator/console-operator-58897d9998-5wrws" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.772759 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/586fee81-a26a-4506-8dfa-66db12798618-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-bnvc4\" (UID: \"586fee81-a26a-4506-8dfa-66db12798618\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bnvc4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.773105 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ad155aa0-50a3-4874-93bd-2166a8d093cb-console-oauth-config\") pod \"console-f9d7485db-f44r4\" (UID: \"ad155aa0-50a3-4874-93bd-2166a8d093cb\") " pod="openshift-console/console-f9d7485db-f44r4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.773369 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/3978218d-91e2-4b20-8bcb-d1d77f820d88-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-6zlqv\" (UID: \"3978218d-91e2-4b20-8bcb-d1d77f820d88\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6zlqv" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.773439 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8391ee1a-9c57-4745-825f-381f47254807-stats-auth\") pod \"router-default-5444994796-wjz5g\" (UID: \"8391ee1a-9c57-4745-825f-381f47254807\") " pod="openshift-ingress/router-default-5444994796-wjz5g" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.774555 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.775204 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.774480 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fbf54dc7-2324-4259-b987-a0df48e06da5-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7qgvq\" (UID: \"fbf54dc7-2324-4259-b987-a0df48e06da5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7qgvq" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.776735 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/39e0228e-f4f6-4167-9bbc-d319374345a1-installation-pull-secrets\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.776942 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da6bd676-392c-4a8a-8a05-dee3304ef594-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-pxlv9\" (UID: \"da6bd676-392c-4a8a-8a05-dee3304ef594\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxlv9" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.777023 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6d137db5-66a8-4066-b8d2-e88adcaaf12c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-66hrr\" (UID: \"6d137db5-66a8-4066-b8d2-e88adcaaf12c\") " pod="openshift-marketplace/marketplace-operator-79b997595-66hrr" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.778330 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8391ee1a-9c57-4745-825f-381f47254807-default-certificate\") pod \"router-default-5444994796-wjz5g\" (UID: \"8391ee1a-9c57-4745-825f-381f47254807\") " pod="openshift-ingress/router-default-5444994796-wjz5g" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.781214 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/586fee81-a26a-4506-8dfa-66db12798618-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-bnvc4\" (UID: \"586fee81-a26a-4506-8dfa-66db12798618\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bnvc4" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.802675 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3978218d-91e2-4b20-8bcb-d1d77f820d88-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-6zlqv\" (UID: \"3978218d-91e2-4b20-8bcb-d1d77f820d88\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6zlqv" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.821219 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/39e0228e-f4f6-4167-9bbc-d319374345a1-bound-sa-token\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.838928 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-947cc\" (UniqueName: \"kubernetes.io/projected/8391ee1a-9c57-4745-825f-381f47254807-kube-api-access-947cc\") pod \"router-default-5444994796-wjz5g\" (UID: \"8391ee1a-9c57-4745-825f-381f47254807\") " pod="openshift-ingress/router-default-5444994796-wjz5g" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.862144 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdxxv\" (UniqueName: \"kubernetes.io/projected/aa8f023f-df93-4883-9771-b0b23c472246-kube-api-access-rdxxv\") pod \"kube-storage-version-migrator-operator-b67b599dd-hb67r\" (UID: \"aa8f023f-df93-4883-9771-b0b23c472246\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb67r" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.866439 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:47 crc kubenswrapper[4934]: E1002 09:50:47.866567 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:48.366533269 +0000 UTC m=+120.119174791 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.866746 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmxrj\" (UniqueName: \"kubernetes.io/projected/9e12b431-672f-404d-99ba-de5f718f19f2-kube-api-access-zmxrj\") pod \"etcd-operator-b45778765-86wbg\" (UID: \"9e12b431-672f-404d-99ba-de5f718f19f2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-86wbg" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.866771 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bszcl\" (UniqueName: \"kubernetes.io/projected/305740e5-7461-4097-994e-731594c096bb-kube-api-access-bszcl\") pod \"machine-config-operator-74547568cd-7frl9\" (UID: \"305740e5-7461-4097-994e-731594c096bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7frl9" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.866810 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/327b634c-c056-49ff-a0ad-a178bc67ac48-apiservice-cert\") pod \"packageserver-d55dfcdfc-wlf9l\" (UID: \"327b634c-c056-49ff-a0ad-a178bc67ac48\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wlf9l" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.866835 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/f4baf828-a5bb-4d7e-a43d-1194d47e4051-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5s688\" (UID: \"f4baf828-a5bb-4d7e-a43d-1194d47e4051\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5s688" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.866858 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5q62x\" (UniqueName: \"kubernetes.io/projected/4b619ff7-c8d0-43ea-9373-18c36b2e160e-kube-api-access-5q62x\") pod \"authentication-operator-69f744f599-llqjv\" (UID: \"4b619ff7-c8d0-43ea-9373-18c36b2e160e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-llqjv" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.866876 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e12b431-672f-404d-99ba-de5f718f19f2-serving-cert\") pod \"etcd-operator-b45778765-86wbg\" (UID: \"9e12b431-672f-404d-99ba-de5f718f19f2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-86wbg" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.866928 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tb84s\" (UniqueName: \"kubernetes.io/projected/90d8cb96-af2e-41b8-a405-85cf9e017631-kube-api-access-tb84s\") pod \"multus-admission-controller-857f4d67dd-hr945\" (UID: \"90d8cb96-af2e-41b8-a405-85cf9e017631\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hr945" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.866950 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e12b431-672f-404d-99ba-de5f718f19f2-config\") pod \"etcd-operator-b45778765-86wbg\" (UID: \"9e12b431-672f-404d-99ba-de5f718f19f2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-86wbg" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.866971 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/9e12b431-672f-404d-99ba-de5f718f19f2-etcd-ca\") pod \"etcd-operator-b45778765-86wbg\" (UID: \"9e12b431-672f-404d-99ba-de5f718f19f2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-86wbg" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.866991 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jfwf\" (UniqueName: \"kubernetes.io/projected/59279e86-a90f-463e-a098-da83a2bdf5a5-kube-api-access-7jfwf\") pod \"machine-config-controller-84d6567774-6jtqh\" (UID: \"59279e86-a90f-463e-a098-da83a2bdf5a5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6jtqh" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.867011 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l46wc\" (UniqueName: \"kubernetes.io/projected/dcb231bb-d375-49d3-b967-dc739da957fb-kube-api-access-l46wc\") pod \"dns-default-v58f6\" (UID: \"dcb231bb-d375-49d3-b967-dc739da957fb\") " pod="openshift-dns/dns-default-v58f6" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.867030 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/327b634c-c056-49ff-a0ad-a178bc67ac48-tmpfs\") pod \"packageserver-d55dfcdfc-wlf9l\" (UID: \"327b634c-c056-49ff-a0ad-a178bc67ac48\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wlf9l" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.867051 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/de99fcca-2d10-4595-8827-410ac05c5f68-registration-dir\") pod \"csi-hostpathplugin-pfl7n\" (UID: \"de99fcca-2d10-4595-8827-410ac05c5f68\") " pod="hostpath-provisioner/csi-hostpathplugin-pfl7n" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.867073 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dcb231bb-d375-49d3-b967-dc739da957fb-config-volume\") pod \"dns-default-v58f6\" (UID: \"dcb231bb-d375-49d3-b967-dc739da957fb\") " pod="openshift-dns/dns-default-v58f6" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.867096 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b52j8\" (UniqueName: \"kubernetes.io/projected/ef34af0b-654c-4519-9292-4e1a833b2ed3-kube-api-access-b52j8\") pod \"control-plane-machine-set-operator-78cbb6b69f-z2d6w\" (UID: \"ef34af0b-654c-4519-9292-4e1a833b2ed3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z2d6w" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.867116 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrp8h\" (UniqueName: \"kubernetes.io/projected/66183279-91fd-423e-91bd-19b68dda8ef0-kube-api-access-lrp8h\") pod \"catalog-operator-68c6474976-9ts9g\" (UID: \"66183279-91fd-423e-91bd-19b68dda8ef0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9ts9g" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.867136 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/25f4d147-18a5-4d67-99c5-39273eecacc0-cert\") pod \"ingress-canary-pzrnw\" (UID: \"25f4d147-18a5-4d67-99c5-39273eecacc0\") " pod="openshift-ingress-canary/ingress-canary-pzrnw" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.867160 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfzlb\" (UniqueName: \"kubernetes.io/projected/8549da74-9ee6-46ff-a0dc-5b4af7d6cfa6-kube-api-access-lfzlb\") pod \"service-ca-operator-777779d784-hw97b\" (UID: \"8549da74-9ee6-46ff-a0dc-5b4af7d6cfa6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hw97b" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.867185 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/90d8cb96-af2e-41b8-a405-85cf9e017631-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-hr945\" (UID: \"90d8cb96-af2e-41b8-a405-85cf9e017631\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hr945" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.867203 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pshs9\" (UniqueName: \"kubernetes.io/projected/ad1a290c-1e04-46fe-8118-92bab2478385-kube-api-access-pshs9\") pod \"ingress-operator-5b745b69d9-x7mrm\" (UID: \"ad1a290c-1e04-46fe-8118-92bab2478385\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-x7mrm" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.867218 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/9e12b431-672f-404d-99ba-de5f718f19f2-etcd-service-ca\") pod \"etcd-operator-b45778765-86wbg\" (UID: \"9e12b431-672f-404d-99ba-de5f718f19f2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-86wbg" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.867232 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/66183279-91fd-423e-91bd-19b68dda8ef0-srv-cert\") pod \"catalog-operator-68c6474976-9ts9g\" (UID: \"66183279-91fd-423e-91bd-19b68dda8ef0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9ts9g" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.867250 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/88ecaab6-9c13-4813-8698-5b558d32a2e9-srv-cert\") pod \"olm-operator-6b444d44fb-lch8r\" (UID: \"88ecaab6-9c13-4813-8698-5b558d32a2e9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lch8r" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.867265 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a-secret-volume\") pod \"collect-profiles-29323305-jv64v\" (UID: \"c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-jv64v" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.867288 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrhkt\" (UniqueName: \"kubernetes.io/projected/25f4d147-18a5-4d67-99c5-39273eecacc0-kube-api-access-vrhkt\") pod \"ingress-canary-pzrnw\" (UID: \"25f4d147-18a5-4d67-99c5-39273eecacc0\") " pod="openshift-ingress-canary/ingress-canary-pzrnw" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.867305 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsgsr\" (UniqueName: \"kubernetes.io/projected/c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a-kube-api-access-vsgsr\") pod \"collect-profiles-29323305-jv64v\" (UID: \"c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-jv64v" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.867322 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/88ecaab6-9c13-4813-8698-5b558d32a2e9-profile-collector-cert\") pod \"olm-operator-6b444d44fb-lch8r\" (UID: \"88ecaab6-9c13-4813-8698-5b558d32a2e9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lch8r" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.867337 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/9fd0eab0-2348-495b-83cf-580fedb4c1f1-certs\") pod \"machine-config-server-w6z57\" (UID: \"9fd0eab0-2348-495b-83cf-580fedb4c1f1\") " pod="openshift-machine-config-operator/machine-config-server-w6z57" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.867445 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/327b634c-c056-49ff-a0ad-a178bc67ac48-webhook-cert\") pod \"packageserver-d55dfcdfc-wlf9l\" (UID: \"327b634c-c056-49ff-a0ad-a178bc67ac48\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wlf9l" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.867481 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7kk9\" (UniqueName: \"kubernetes.io/projected/28aceb4d-e776-4340-9812-8150b2766d5c-kube-api-access-r7kk9\") pod \"migrator-59844c95c7-8xjjw\" (UID: \"28aceb4d-e776-4340-9812-8150b2766d5c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8xjjw" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.867505 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b619ff7-c8d0-43ea-9373-18c36b2e160e-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-llqjv\" (UID: \"4b619ff7-c8d0-43ea-9373-18c36b2e160e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-llqjv" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.867525 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ad1a290c-1e04-46fe-8118-92bab2478385-bound-sa-token\") pod \"ingress-operator-5b745b69d9-x7mrm\" (UID: \"ad1a290c-1e04-46fe-8118-92bab2478385\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-x7mrm" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.867549 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/59279e86-a90f-463e-a098-da83a2bdf5a5-proxy-tls\") pod \"machine-config-controller-84d6567774-6jtqh\" (UID: \"59279e86-a90f-463e-a098-da83a2bdf5a5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6jtqh" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.867575 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a-config-volume\") pod \"collect-profiles-29323305-jv64v\" (UID: \"c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-jv64v" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.867620 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8549da74-9ee6-46ff-a0dc-5b4af7d6cfa6-config\") pod \"service-ca-operator-777779d784-hw97b\" (UID: \"8549da74-9ee6-46ff-a0dc-5b4af7d6cfa6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hw97b" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.867644 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/305740e5-7461-4097-994e-731594c096bb-auth-proxy-config\") pod \"machine-config-operator-74547568cd-7frl9\" (UID: \"305740e5-7461-4097-994e-731594c096bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7frl9" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.867663 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxcmq\" (UniqueName: \"kubernetes.io/projected/8f6c0ab1-dd58-4734-89f8-aefb606078d3-kube-api-access-pxcmq\") pod \"service-ca-9c57cc56f-2vn7r\" (UID: \"8f6c0ab1-dd58-4734-89f8-aefb606078d3\") " pod="openshift-service-ca/service-ca-9c57cc56f-2vn7r" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.868719 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/66183279-91fd-423e-91bd-19b68dda8ef0-profile-collector-cert\") pod \"catalog-operator-68c6474976-9ts9g\" (UID: \"66183279-91fd-423e-91bd-19b68dda8ef0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9ts9g" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.868784 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b619ff7-c8d0-43ea-9373-18c36b2e160e-config\") pod \"authentication-operator-69f744f599-llqjv\" (UID: \"4b619ff7-c8d0-43ea-9373-18c36b2e160e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-llqjv" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869047 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bd4h\" (UniqueName: \"kubernetes.io/projected/88ecaab6-9c13-4813-8698-5b558d32a2e9-kube-api-access-4bd4h\") pod \"olm-operator-6b444d44fb-lch8r\" (UID: \"88ecaab6-9c13-4813-8698-5b558d32a2e9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lch8r" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869106 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/de99fcca-2d10-4595-8827-410ac05c5f68-mountpoint-dir\") pod \"csi-hostpathplugin-pfl7n\" (UID: \"de99fcca-2d10-4595-8827-410ac05c5f68\") " pod="hostpath-provisioner/csi-hostpathplugin-pfl7n" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869134 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869167 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/de99fcca-2d10-4595-8827-410ac05c5f68-socket-dir\") pod \"csi-hostpathplugin-pfl7n\" (UID: \"de99fcca-2d10-4595-8827-410ac05c5f68\") " pod="hostpath-provisioner/csi-hostpathplugin-pfl7n" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869190 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/8f6c0ab1-dd58-4734-89f8-aefb606078d3-signing-cabundle\") pod \"service-ca-9c57cc56f-2vn7r\" (UID: \"8f6c0ab1-dd58-4734-89f8-aefb606078d3\") " pod="openshift-service-ca/service-ca-9c57cc56f-2vn7r" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869214 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrqkv\" (UniqueName: \"kubernetes.io/projected/f4baf828-a5bb-4d7e-a43d-1194d47e4051-kube-api-access-jrqkv\") pod \"package-server-manager-789f6589d5-5s688\" (UID: \"f4baf828-a5bb-4d7e-a43d-1194d47e4051\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5s688" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869234 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e12b431-672f-404d-99ba-de5f718f19f2-config\") pod \"etcd-operator-b45778765-86wbg\" (UID: \"9e12b431-672f-404d-99ba-de5f718f19f2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-86wbg" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869283 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/de99fcca-2d10-4595-8827-410ac05c5f68-registration-dir\") pod \"csi-hostpathplugin-pfl7n\" (UID: \"de99fcca-2d10-4595-8827-410ac05c5f68\") " pod="hostpath-provisioner/csi-hostpathplugin-pfl7n" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869246 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b619ff7-c8d0-43ea-9373-18c36b2e160e-serving-cert\") pod \"authentication-operator-69f744f599-llqjv\" (UID: \"4b619ff7-c8d0-43ea-9373-18c36b2e160e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-llqjv" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869346 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/ef34af0b-654c-4519-9292-4e1a833b2ed3-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-z2d6w\" (UID: \"ef34af0b-654c-4519-9292-4e1a833b2ed3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z2d6w" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869376 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbqk7\" (UniqueName: \"kubernetes.io/projected/9fd0eab0-2348-495b-83cf-580fedb4c1f1-kube-api-access-zbqk7\") pod \"machine-config-server-w6z57\" (UID: \"9fd0eab0-2348-495b-83cf-580fedb4c1f1\") " pod="openshift-machine-config-operator/machine-config-server-w6z57" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869403 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/59279e86-a90f-463e-a098-da83a2bdf5a5-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-6jtqh\" (UID: \"59279e86-a90f-463e-a098-da83a2bdf5a5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6jtqh" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869426 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/305740e5-7461-4097-994e-731594c096bb-proxy-tls\") pod \"machine-config-operator-74547568cd-7frl9\" (UID: \"305740e5-7461-4097-994e-731594c096bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7frl9" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869463 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jflw\" (UniqueName: \"kubernetes.io/projected/327b634c-c056-49ff-a0ad-a178bc67ac48-kube-api-access-6jflw\") pod \"packageserver-d55dfcdfc-wlf9l\" (UID: \"327b634c-c056-49ff-a0ad-a178bc67ac48\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wlf9l" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869494 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ad1a290c-1e04-46fe-8118-92bab2478385-metrics-tls\") pod \"ingress-operator-5b745b69d9-x7mrm\" (UID: \"ad1a290c-1e04-46fe-8118-92bab2478385\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-x7mrm" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869515 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ad1a290c-1e04-46fe-8118-92bab2478385-trusted-ca\") pod \"ingress-operator-5b745b69d9-x7mrm\" (UID: \"ad1a290c-1e04-46fe-8118-92bab2478385\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-x7mrm" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869542 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b619ff7-c8d0-43ea-9373-18c36b2e160e-service-ca-bundle\") pod \"authentication-operator-69f744f599-llqjv\" (UID: \"4b619ff7-c8d0-43ea-9373-18c36b2e160e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-llqjv" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869566 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/9fd0eab0-2348-495b-83cf-580fedb4c1f1-node-bootstrap-token\") pod \"machine-config-server-w6z57\" (UID: \"9fd0eab0-2348-495b-83cf-580fedb4c1f1\") " pod="openshift-machine-config-operator/machine-config-server-w6z57" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869614 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/8f6c0ab1-dd58-4734-89f8-aefb606078d3-signing-key\") pod \"service-ca-9c57cc56f-2vn7r\" (UID: \"8f6c0ab1-dd58-4734-89f8-aefb606078d3\") " pod="openshift-service-ca/service-ca-9c57cc56f-2vn7r" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869638 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8549da74-9ee6-46ff-a0dc-5b4af7d6cfa6-serving-cert\") pod \"service-ca-operator-777779d784-hw97b\" (UID: \"8549da74-9ee6-46ff-a0dc-5b4af7d6cfa6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hw97b" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869660 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/305740e5-7461-4097-994e-731594c096bb-images\") pod \"machine-config-operator-74547568cd-7frl9\" (UID: \"305740e5-7461-4097-994e-731594c096bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7frl9" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869682 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/dcb231bb-d375-49d3-b967-dc739da957fb-metrics-tls\") pod \"dns-default-v58f6\" (UID: \"dcb231bb-d375-49d3-b967-dc739da957fb\") " pod="openshift-dns/dns-default-v58f6" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869722 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/de99fcca-2d10-4595-8827-410ac05c5f68-csi-data-dir\") pod \"csi-hostpathplugin-pfl7n\" (UID: \"de99fcca-2d10-4595-8827-410ac05c5f68\") " pod="hostpath-provisioner/csi-hostpathplugin-pfl7n" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869744 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mll7c\" (UniqueName: \"kubernetes.io/projected/de99fcca-2d10-4595-8827-410ac05c5f68-kube-api-access-mll7c\") pod \"csi-hostpathplugin-pfl7n\" (UID: \"de99fcca-2d10-4595-8827-410ac05c5f68\") " pod="hostpath-provisioner/csi-hostpathplugin-pfl7n" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869765 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/de99fcca-2d10-4595-8827-410ac05c5f68-plugins-dir\") pod \"csi-hostpathplugin-pfl7n\" (UID: \"de99fcca-2d10-4595-8827-410ac05c5f68\") " pod="hostpath-provisioner/csi-hostpathplugin-pfl7n" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869789 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9e12b431-672f-404d-99ba-de5f718f19f2-etcd-client\") pod \"etcd-operator-b45778765-86wbg\" (UID: \"9e12b431-672f-404d-99ba-de5f718f19f2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-86wbg" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869796 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/305740e5-7461-4097-994e-731594c096bb-auth-proxy-config\") pod \"machine-config-operator-74547568cd-7frl9\" (UID: \"305740e5-7461-4097-994e-731594c096bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7frl9" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869875 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a-config-volume\") pod \"collect-profiles-29323305-jv64v\" (UID: \"c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-jv64v" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.869898 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b619ff7-c8d0-43ea-9373-18c36b2e160e-config\") pod \"authentication-operator-69f744f599-llqjv\" (UID: \"4b619ff7-c8d0-43ea-9373-18c36b2e160e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-llqjv" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.870153 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/327b634c-c056-49ff-a0ad-a178bc67ac48-tmpfs\") pod \"packageserver-d55dfcdfc-wlf9l\" (UID: \"327b634c-c056-49ff-a0ad-a178bc67ac48\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wlf9l" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.870388 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b619ff7-c8d0-43ea-9373-18c36b2e160e-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-llqjv\" (UID: \"4b619ff7-c8d0-43ea-9373-18c36b2e160e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-llqjv" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.870461 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8549da74-9ee6-46ff-a0dc-5b4af7d6cfa6-config\") pod \"service-ca-operator-777779d784-hw97b\" (UID: \"8549da74-9ee6-46ff-a0dc-5b4af7d6cfa6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hw97b" Oct 02 09:50:47 crc kubenswrapper[4934]: E1002 09:50:47.871207 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:48.371193466 +0000 UTC m=+120.123835088 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.871622 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/de99fcca-2d10-4595-8827-410ac05c5f68-socket-dir\") pod \"csi-hostpathplugin-pfl7n\" (UID: \"de99fcca-2d10-4595-8827-410ac05c5f68\") " pod="hostpath-provisioner/csi-hostpathplugin-pfl7n" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.872264 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dcb231bb-d375-49d3-b967-dc739da957fb-config-volume\") pod \"dns-default-v58f6\" (UID: \"dcb231bb-d375-49d3-b967-dc739da957fb\") " pod="openshift-dns/dns-default-v58f6" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.872410 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/de99fcca-2d10-4595-8827-410ac05c5f68-mountpoint-dir\") pod \"csi-hostpathplugin-pfl7n\" (UID: \"de99fcca-2d10-4595-8827-410ac05c5f68\") " pod="hostpath-provisioner/csi-hostpathplugin-pfl7n" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.872998 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/de99fcca-2d10-4595-8827-410ac05c5f68-csi-data-dir\") pod \"csi-hostpathplugin-pfl7n\" (UID: \"de99fcca-2d10-4595-8827-410ac05c5f68\") " pod="hostpath-provisioner/csi-hostpathplugin-pfl7n" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.873531 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b619ff7-c8d0-43ea-9373-18c36b2e160e-service-ca-bundle\") pod \"authentication-operator-69f744f599-llqjv\" (UID: \"4b619ff7-c8d0-43ea-9373-18c36b2e160e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-llqjv" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.873606 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/9e12b431-672f-404d-99ba-de5f718f19f2-etcd-ca\") pod \"etcd-operator-b45778765-86wbg\" (UID: \"9e12b431-672f-404d-99ba-de5f718f19f2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-86wbg" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.873689 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/9e12b431-672f-404d-99ba-de5f718f19f2-etcd-service-ca\") pod \"etcd-operator-b45778765-86wbg\" (UID: \"9e12b431-672f-404d-99ba-de5f718f19f2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-86wbg" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.874048 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/de99fcca-2d10-4595-8827-410ac05c5f68-plugins-dir\") pod \"csi-hostpathplugin-pfl7n\" (UID: \"de99fcca-2d10-4595-8827-410ac05c5f68\") " pod="hostpath-provisioner/csi-hostpathplugin-pfl7n" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.874347 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/59279e86-a90f-463e-a098-da83a2bdf5a5-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-6jtqh\" (UID: \"59279e86-a90f-463e-a098-da83a2bdf5a5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6jtqh" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.874694 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/8f6c0ab1-dd58-4734-89f8-aefb606078d3-signing-cabundle\") pod \"service-ca-9c57cc56f-2vn7r\" (UID: \"8f6c0ab1-dd58-4734-89f8-aefb606078d3\") " pod="openshift-service-ca/service-ca-9c57cc56f-2vn7r" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.875727 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e12b431-672f-404d-99ba-de5f718f19f2-serving-cert\") pod \"etcd-operator-b45778765-86wbg\" (UID: \"9e12b431-672f-404d-99ba-de5f718f19f2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-86wbg" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.876358 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a-secret-volume\") pod \"collect-profiles-29323305-jv64v\" (UID: \"c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-jv64v" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.877136 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b619ff7-c8d0-43ea-9373-18c36b2e160e-serving-cert\") pod \"authentication-operator-69f744f599-llqjv\" (UID: \"4b619ff7-c8d0-43ea-9373-18c36b2e160e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-llqjv" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.877522 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/25f4d147-18a5-4d67-99c5-39273eecacc0-cert\") pod \"ingress-canary-pzrnw\" (UID: \"25f4d147-18a5-4d67-99c5-39273eecacc0\") " pod="openshift-ingress-canary/ingress-canary-pzrnw" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.877548 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/327b634c-c056-49ff-a0ad-a178bc67ac48-apiservice-cert\") pod \"packageserver-d55dfcdfc-wlf9l\" (UID: \"327b634c-c056-49ff-a0ad-a178bc67ac48\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wlf9l" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.877545 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ad1a290c-1e04-46fe-8118-92bab2478385-trusted-ca\") pod \"ingress-operator-5b745b69d9-x7mrm\" (UID: \"ad1a290c-1e04-46fe-8118-92bab2478385\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-x7mrm" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.877767 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/9fd0eab0-2348-495b-83cf-580fedb4c1f1-certs\") pod \"machine-config-server-w6z57\" (UID: \"9fd0eab0-2348-495b-83cf-580fedb4c1f1\") " pod="openshift-machine-config-operator/machine-config-server-w6z57" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.877802 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/305740e5-7461-4097-994e-731594c096bb-images\") pod \"machine-config-operator-74547568cd-7frl9\" (UID: \"305740e5-7461-4097-994e-731594c096bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7frl9" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.878030 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9e12b431-672f-404d-99ba-de5f718f19f2-etcd-client\") pod \"etcd-operator-b45778765-86wbg\" (UID: \"9e12b431-672f-404d-99ba-de5f718f19f2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-86wbg" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.878725 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/dcb231bb-d375-49d3-b967-dc739da957fb-metrics-tls\") pod \"dns-default-v58f6\" (UID: \"dcb231bb-d375-49d3-b967-dc739da957fb\") " pod="openshift-dns/dns-default-v58f6" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.879506 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/66183279-91fd-423e-91bd-19b68dda8ef0-srv-cert\") pod \"catalog-operator-68c6474976-9ts9g\" (UID: \"66183279-91fd-423e-91bd-19b68dda8ef0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9ts9g" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.879566 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/f4baf828-a5bb-4d7e-a43d-1194d47e4051-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5s688\" (UID: \"f4baf828-a5bb-4d7e-a43d-1194d47e4051\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5s688" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.879596 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/66183279-91fd-423e-91bd-19b68dda8ef0-profile-collector-cert\") pod \"catalog-operator-68c6474976-9ts9g\" (UID: \"66183279-91fd-423e-91bd-19b68dda8ef0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9ts9g" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.879745 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/305740e5-7461-4097-994e-731594c096bb-proxy-tls\") pod \"machine-config-operator-74547568cd-7frl9\" (UID: \"305740e5-7461-4097-994e-731594c096bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7frl9" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.879981 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8549da74-9ee6-46ff-a0dc-5b4af7d6cfa6-serving-cert\") pod \"service-ca-operator-777779d784-hw97b\" (UID: \"8549da74-9ee6-46ff-a0dc-5b4af7d6cfa6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hw97b" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.880242 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/88ecaab6-9c13-4813-8698-5b558d32a2e9-srv-cert\") pod \"olm-operator-6b444d44fb-lch8r\" (UID: \"88ecaab6-9c13-4813-8698-5b558d32a2e9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lch8r" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.880965 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/9fd0eab0-2348-495b-83cf-580fedb4c1f1-node-bootstrap-token\") pod \"machine-config-server-w6z57\" (UID: \"9fd0eab0-2348-495b-83cf-580fedb4c1f1\") " pod="openshift-machine-config-operator/machine-config-server-w6z57" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.882088 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/59279e86-a90f-463e-a098-da83a2bdf5a5-proxy-tls\") pod \"machine-config-controller-84d6567774-6jtqh\" (UID: \"59279e86-a90f-463e-a098-da83a2bdf5a5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6jtqh" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.883322 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/88ecaab6-9c13-4813-8698-5b558d32a2e9-profile-collector-cert\") pod \"olm-operator-6b444d44fb-lch8r\" (UID: \"88ecaab6-9c13-4813-8698-5b558d32a2e9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lch8r" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.883368 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/8f6c0ab1-dd58-4734-89f8-aefb606078d3-signing-key\") pod \"service-ca-9c57cc56f-2vn7r\" (UID: \"8f6c0ab1-dd58-4734-89f8-aefb606078d3\") " pod="openshift-service-ca/service-ca-9c57cc56f-2vn7r" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.883796 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/90d8cb96-af2e-41b8-a405-85cf9e017631-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-hr945\" (UID: \"90d8cb96-af2e-41b8-a405-85cf9e017631\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hr945" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.887967 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/327b634c-c056-49ff-a0ad-a178bc67ac48-webhook-cert\") pod \"packageserver-d55dfcdfc-wlf9l\" (UID: \"327b634c-c056-49ff-a0ad-a178bc67ac48\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wlf9l" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.888031 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ad1a290c-1e04-46fe-8118-92bab2478385-metrics-tls\") pod \"ingress-operator-5b745b69d9-x7mrm\" (UID: \"ad1a290c-1e04-46fe-8118-92bab2478385\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-x7mrm" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.888091 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/ef34af0b-654c-4519-9292-4e1a833b2ed3-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-z2d6w\" (UID: \"ef34af0b-654c-4519-9292-4e1a833b2ed3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z2d6w" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.888236 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d88b\" (UniqueName: \"kubernetes.io/projected/5dbb4e0e-9898-44ab-81f9-d149b80cfef0-kube-api-access-8d88b\") pod \"downloads-7954f5f757-mcnd2\" (UID: \"5dbb4e0e-9898-44ab-81f9-d149b80cfef0\") " pod="openshift-console/downloads-7954f5f757-mcnd2" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.907037 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66c8v\" (UniqueName: \"kubernetes.io/projected/da6bd676-392c-4a8a-8a05-dee3304ef594-kube-api-access-66c8v\") pod \"openshift-apiserver-operator-796bbdcf4f-pxlv9\" (UID: \"da6bd676-392c-4a8a-8a05-dee3304ef594\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxlv9" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.922291 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55tcl\" (UniqueName: \"kubernetes.io/projected/3978218d-91e2-4b20-8bcb-d1d77f820d88-kube-api-access-55tcl\") pod \"cluster-image-registry-operator-dc59b4c8b-6zlqv\" (UID: \"3978218d-91e2-4b20-8bcb-d1d77f820d88\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6zlqv" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.938738 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6zlqv" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.945929 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-mcnd2" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.952065 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fbf54dc7-2324-4259-b987-a0df48e06da5-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7qgvq\" (UID: \"fbf54dc7-2324-4259-b987-a0df48e06da5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7qgvq" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.962915 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ebd4ee43-4dc5-46dc-964b-236cd1b16ec4-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-qg98s\" (UID: \"ebd4ee43-4dc5-46dc-964b-236cd1b16ec4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qg98s" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.971482 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:47 crc kubenswrapper[4934]: E1002 09:50:47.971697 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:48.471666134 +0000 UTC m=+120.224307656 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.971876 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:47 crc kubenswrapper[4934]: E1002 09:50:47.973149 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:48.473127189 +0000 UTC m=+120.225768811 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.992992 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qg98s" Oct 02 09:50:47 crc kubenswrapper[4934]: I1002 09:50:47.998853 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb67r" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.001909 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t857q\" (UniqueName: \"kubernetes.io/projected/460de405-cddd-4e59-8be4-537419c25f55-kube-api-access-t857q\") pod \"dns-operator-744455d44c-rjhkx\" (UID: \"460de405-cddd-4e59-8be4-537419c25f55\") " pod="openshift-dns-operator/dns-operator-744455d44c-rjhkx" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.009052 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7qgvq" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.016038 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-wjz5g" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.017718 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxlv9" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.024717 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bnvc4" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.025747 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvmvq\" (UniqueName: \"kubernetes.io/projected/39e0228e-f4f6-4167-9bbc-d319374345a1-kube-api-access-fvmvq\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.051449 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l42zk\" (UniqueName: \"kubernetes.io/projected/6d137db5-66a8-4066-b8d2-e88adcaaf12c-kube-api-access-l42zk\") pod \"marketplace-operator-79b997595-66hrr\" (UID: \"6d137db5-66a8-4066-b8d2-e88adcaaf12c\") " pod="openshift-marketplace/marketplace-operator-79b997595-66hrr" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.067153 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kr89h\" (UniqueName: \"kubernetes.io/projected/ad155aa0-50a3-4874-93bd-2166a8d093cb-kube-api-access-kr89h\") pod \"console-f9d7485db-f44r4\" (UID: \"ad155aa0-50a3-4874-93bd-2166a8d093cb\") " pod="openshift-console/console-f9d7485db-f44r4" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.074005 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:48 crc kubenswrapper[4934]: E1002 09:50:48.074459 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:48.574442717 +0000 UTC m=+120.327084239 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.085521 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b47xn\" (UniqueName: \"kubernetes.io/projected/5f7c01ef-b9b0-476b-b3e0-12b8132b58b9-kube-api-access-b47xn\") pod \"console-operator-58897d9998-5wrws\" (UID: \"5f7c01ef-b9b0-476b-b3e0-12b8132b58b9\") " pod="openshift-console-operator/console-operator-58897d9998-5wrws" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.110314 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc58m\" (UniqueName: \"kubernetes.io/projected/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-kube-api-access-rc58m\") pod \"oauth-openshift-558db77b4-4pp8p\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.124373 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrbrd\" (UniqueName: \"kubernetes.io/projected/fec65034-065a-466e-8c61-dacc1e415b6f-kube-api-access-vrbrd\") pod \"openshift-controller-manager-operator-756b6f6bc6-7t8rp\" (UID: \"fec65034-065a-466e-8c61-dacc1e415b6f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7t8rp" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.146362 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jfwf\" (UniqueName: \"kubernetes.io/projected/59279e86-a90f-463e-a098-da83a2bdf5a5-kube-api-access-7jfwf\") pod \"machine-config-controller-84d6567774-6jtqh\" (UID: \"59279e86-a90f-463e-a098-da83a2bdf5a5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6jtqh" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.174390 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6zlqv"] Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.174862 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:48 crc kubenswrapper[4934]: E1002 09:50:48.175233 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:48.675218622 +0000 UTC m=+120.427860144 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.189740 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmxrj\" (UniqueName: \"kubernetes.io/projected/9e12b431-672f-404d-99ba-de5f718f19f2-kube-api-access-zmxrj\") pod \"etcd-operator-b45778765-86wbg\" (UID: \"9e12b431-672f-404d-99ba-de5f718f19f2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-86wbg" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.193102 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bszcl\" (UniqueName: \"kubernetes.io/projected/305740e5-7461-4097-994e-731594c096bb-kube-api-access-bszcl\") pod \"machine-config-operator-74547568cd-7frl9\" (UID: \"305740e5-7461-4097-994e-731594c096bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7frl9" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.197686 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-f44r4" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.205913 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7kk9\" (UniqueName: \"kubernetes.io/projected/28aceb4d-e776-4340-9812-8150b2766d5c-kube-api-access-r7kk9\") pod \"migrator-59844c95c7-8xjjw\" (UID: \"28aceb4d-e776-4340-9812-8150b2766d5c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8xjjw" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.206222 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7t8rp" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.211525 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-mcnd2"] Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.224487 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-rjhkx" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.228271 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l46wc\" (UniqueName: \"kubernetes.io/projected/dcb231bb-d375-49d3-b967-dc739da957fb-kube-api-access-l46wc\") pod \"dns-default-v58f6\" (UID: \"dcb231bb-d375-49d3-b967-dc739da957fb\") " pod="openshift-dns/dns-default-v58f6" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.245254 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrhkt\" (UniqueName: \"kubernetes.io/projected/25f4d147-18a5-4d67-99c5-39273eecacc0-kube-api-access-vrhkt\") pod \"ingress-canary-pzrnw\" (UID: \"25f4d147-18a5-4d67-99c5-39273eecacc0\") " pod="openshift-ingress-canary/ingress-canary-pzrnw" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.256937 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.261546 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrp8h\" (UniqueName: \"kubernetes.io/projected/66183279-91fd-423e-91bd-19b68dda8ef0-kube-api-access-lrp8h\") pod \"catalog-operator-68c6474976-9ts9g\" (UID: \"66183279-91fd-423e-91bd-19b68dda8ef0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9ts9g" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.275125 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:48 crc kubenswrapper[4934]: E1002 09:50:48.275681 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:48.77566072 +0000 UTC m=+120.528302242 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.279364 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-5wrws" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.285453 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tb84s\" (UniqueName: \"kubernetes.io/projected/90d8cb96-af2e-41b8-a405-85cf9e017631-kube-api-access-tb84s\") pod \"multus-admission-controller-857f4d67dd-hr945\" (UID: \"90d8cb96-af2e-41b8-a405-85cf9e017631\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hr945" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.311626 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxcmq\" (UniqueName: \"kubernetes.io/projected/8f6c0ab1-dd58-4734-89f8-aefb606078d3-kube-api-access-pxcmq\") pod \"service-ca-9c57cc56f-2vn7r\" (UID: \"8f6c0ab1-dd58-4734-89f8-aefb606078d3\") " pod="openshift-service-ca/service-ca-9c57cc56f-2vn7r" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.323383 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5q62x\" (UniqueName: \"kubernetes.io/projected/4b619ff7-c8d0-43ea-9373-18c36b2e160e-kube-api-access-5q62x\") pod \"authentication-operator-69f744f599-llqjv\" (UID: \"4b619ff7-c8d0-43ea-9373-18c36b2e160e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-llqjv" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.332709 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-66hrr" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.341294 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7frl9" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.356701 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ad1a290c-1e04-46fe-8118-92bab2478385-bound-sa-token\") pod \"ingress-operator-5b745b69d9-x7mrm\" (UID: \"ad1a290c-1e04-46fe-8118-92bab2478385\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-x7mrm" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.360026 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-86wbg" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.373971 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pshs9\" (UniqueName: \"kubernetes.io/projected/ad1a290c-1e04-46fe-8118-92bab2478385-kube-api-access-pshs9\") pod \"ingress-operator-5b745b69d9-x7mrm\" (UID: \"ad1a290c-1e04-46fe-8118-92bab2478385\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-x7mrm" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.377014 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:48 crc kubenswrapper[4934]: E1002 09:50:48.377386 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:48.877372137 +0000 UTC m=+120.630013659 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.380875 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9ts9g" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.388166 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfzlb\" (UniqueName: \"kubernetes.io/projected/8549da74-9ee6-46ff-a0dc-5b4af7d6cfa6-kube-api-access-lfzlb\") pod \"service-ca-operator-777779d784-hw97b\" (UID: \"8549da74-9ee6-46ff-a0dc-5b4af7d6cfa6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hw97b" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.395933 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6jtqh" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.402747 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hw97b" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.404419 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsgsr\" (UniqueName: \"kubernetes.io/projected/c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a-kube-api-access-vsgsr\") pod \"collect-profiles-29323305-jv64v\" (UID: \"c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-jv64v" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.419976 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-hr945" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.421525 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bd4h\" (UniqueName: \"kubernetes.io/projected/88ecaab6-9c13-4813-8698-5b558d32a2e9-kube-api-access-4bd4h\") pod \"olm-operator-6b444d44fb-lch8r\" (UID: \"88ecaab6-9c13-4813-8698-5b558d32a2e9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lch8r" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.427980 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-pzrnw" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.438005 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8xjjw" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.444326 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrqkv\" (UniqueName: \"kubernetes.io/projected/f4baf828-a5bb-4d7e-a43d-1194d47e4051-kube-api-access-jrqkv\") pod \"package-server-manager-789f6589d5-5s688\" (UID: \"f4baf828-a5bb-4d7e-a43d-1194d47e4051\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5s688" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.450888 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-llqjv" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.463916 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-2vn7r" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.478142 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:48 crc kubenswrapper[4934]: E1002 09:50:48.478523 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:48.978504621 +0000 UTC m=+120.731146153 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.478543 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mll7c\" (UniqueName: \"kubernetes.io/projected/de99fcca-2d10-4595-8827-410ac05c5f68-kube-api-access-mll7c\") pod \"csi-hostpathplugin-pfl7n\" (UID: \"de99fcca-2d10-4595-8827-410ac05c5f68\") " pod="hostpath-provisioner/csi-hostpathplugin-pfl7n" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.479343 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:48 crc kubenswrapper[4934]: E1002 09:50:48.479731 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:48.97972106 +0000 UTC m=+120.732362592 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.486727 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-v58f6" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.488662 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qg98s"] Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.494294 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b52j8\" (UniqueName: \"kubernetes.io/projected/ef34af0b-654c-4519-9292-4e1a833b2ed3-kube-api-access-b52j8\") pod \"control-plane-machine-set-operator-78cbb6b69f-z2d6w\" (UID: \"ef34af0b-654c-4519-9292-4e1a833b2ed3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z2d6w" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.502540 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-pfl7n" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.528852 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbqk7\" (UniqueName: \"kubernetes.io/projected/9fd0eab0-2348-495b-83cf-580fedb4c1f1-kube-api-access-zbqk7\") pod \"machine-config-server-w6z57\" (UID: \"9fd0eab0-2348-495b-83cf-580fedb4c1f1\") " pod="openshift-machine-config-operator/machine-config-server-w6z57" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.536982 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-t6dmx" event={"ID":"8bbc993a-a1ea-48ae-a765-8f656d17329c","Type":"ContainerStarted","Data":"edfa4342231113276fba078e32ea809e972cb9c3ae0133faff032e6df54be057"} Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.537025 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-t6dmx" event={"ID":"8bbc993a-a1ea-48ae-a765-8f656d17329c","Type":"ContainerStarted","Data":"c63ea467fb603d51e4706def19b2d2727b48a9758eb080b8075de02bde49f32d"} Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.537037 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-t6dmx" event={"ID":"8bbc993a-a1ea-48ae-a765-8f656d17329c","Type":"ContainerStarted","Data":"e30dd351f4f91d5e6f7c41828bdcad1259414dd0ec1554032a7d5d82bcab73dd"} Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.541250 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jflw\" (UniqueName: \"kubernetes.io/projected/327b634c-c056-49ff-a0ad-a178bc67ac48-kube-api-access-6jflw\") pod \"packageserver-d55dfcdfc-wlf9l\" (UID: \"327b634c-c056-49ff-a0ad-a178bc67ac48\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wlf9l" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.542515 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-mcnd2" event={"ID":"5dbb4e0e-9898-44ab-81f9-d149b80cfef0","Type":"ContainerStarted","Data":"fe3c4b07aa04eeeacb323aee6fcc84061481f6a87e6d862d8046390e9ce52288"} Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.542572 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-mcnd2" event={"ID":"5dbb4e0e-9898-44ab-81f9-d149b80cfef0","Type":"ContainerStarted","Data":"81edabcbda97c5c5290018b0d29e7a1dca9a6e0644b95b0a1dfc9f73d40a626f"} Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.543616 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-mcnd2" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.555241 4934 patch_prober.go:28] interesting pod/downloads-7954f5f757-mcnd2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.555293 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mcnd2" podUID="5dbb4e0e-9898-44ab-81f9-d149b80cfef0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.556976 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4bqtk" event={"ID":"602dc2aa-bffc-4e88-95ad-5dce0fc6d6df","Type":"ContainerStarted","Data":"2b61aa10d7a20b9b264a15f57fa6c267c86a161db9c146dcc4ba07b74f7ad47a"} Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.557030 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4bqtk" event={"ID":"602dc2aa-bffc-4e88-95ad-5dce0fc6d6df","Type":"ContainerStarted","Data":"20f368274d17086307ace3ea16de8a8d65eb882f3dee9f6e706e57a1e1357ad6"} Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.573519 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bnvc4"] Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.593482 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.597432 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7qgvq"] Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.597476 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb67r"] Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.597487 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxlv9"] Oct 02 09:50:48 crc kubenswrapper[4934]: E1002 09:50:48.599033 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:49.099010652 +0000 UTC m=+120.851652254 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.620248 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6" event={"ID":"87b8c00f-e910-42f0-8089-73df46c6cf0b","Type":"ContainerStarted","Data":"1f42a757d9a24bdf290901c28bdbb637a8e44d2cdd551ec0573b9f29fa2d07f3"} Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.620295 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6" event={"ID":"87b8c00f-e910-42f0-8089-73df46c6cf0b","Type":"ContainerStarted","Data":"20b01acf128a5ec360f96e55587dd495ddd8b1ca26c0f2fe7c6932e2cdd82c4d"} Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.621130 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.627513 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gmxzf" event={"ID":"62e9cfe8-c134-423b-b971-c552f98526f9","Type":"ContainerStarted","Data":"fe40a682606dfca6f50ab63fb62c1682a7b405426731c4eb932ac8960dd511b7"} Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.627542 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gmxzf" event={"ID":"62e9cfe8-c134-423b-b971-c552f98526f9","Type":"ContainerStarted","Data":"d00f9675211bccb2558a93313539334197c51f0d4c0361d55201faf1b8b40b8f"} Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.630038 4934 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-d6cv6 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.630085 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6" podUID="87b8c00f-e910-42f0-8089-73df46c6cf0b" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.630398 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6zlqv" event={"ID":"3978218d-91e2-4b20-8bcb-d1d77f820d88","Type":"ContainerStarted","Data":"08f5b8fd2db8f2f3cdfa020dbd0e0bcc957dfd82611e3cc2af3ab30105260afc"} Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.630428 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6zlqv" event={"ID":"3978218d-91e2-4b20-8bcb-d1d77f820d88","Type":"ContainerStarted","Data":"e137f3d5cb08d3419c67368db0917f885c312b27703f5781eeb9d485af905323"} Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.650360 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-x7mrm" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.659775 4934 generic.go:334] "Generic (PLEG): container finished" podID="1d074a40-447d-4144-a2dc-ca69269c84de" containerID="a02f75ca7a2d96f265bf7595aaf828e981cb9cd4131d6256587074cf06322e30" exitCode=0 Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.659875 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" event={"ID":"1d074a40-447d-4144-a2dc-ca69269c84de","Type":"ContainerDied","Data":"a02f75ca7a2d96f265bf7595aaf828e981cb9cd4131d6256587074cf06322e30"} Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.664262 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7t8rp"] Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.668166 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-jv64v" Oct 02 09:50:48 crc kubenswrapper[4934]: W1002 09:50:48.678814 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfbf54dc7_2324_4259_b987_a0df48e06da5.slice/crio-6030718e1edef2bb249fe8e8f2bea5d2cc502ee1df789bbca3d66b8064a8c5df WatchSource:0}: Error finding container 6030718e1edef2bb249fe8e8f2bea5d2cc502ee1df789bbca3d66b8064a8c5df: Status 404 returned error can't find the container with id 6030718e1edef2bb249fe8e8f2bea5d2cc502ee1df789bbca3d66b8064a8c5df Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.683004 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-wjz5g" event={"ID":"8391ee1a-9c57-4745-825f-381f47254807","Type":"ContainerStarted","Data":"cc7078f4576bb8b9fe6c07102164ba757df86d13cb4f37826e6cca58a5617c7e"} Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.683086 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-wjz5g" event={"ID":"8391ee1a-9c57-4745-825f-381f47254807","Type":"ContainerStarted","Data":"e003e07044000620e1df9f167b9c23e75334078600d5c87e35ca6d7149350dc7"} Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.689448 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5s688" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.697716 4934 generic.go:334] "Generic (PLEG): container finished" podID="6096bf5b-7d9b-41c1-a6b9-d2f6879175a6" containerID="c7a1e801804e99d488b91871797f6061ec5ab018dfd37e42457ab9fe58227712" exitCode=0 Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.698050 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" event={"ID":"6096bf5b-7d9b-41c1-a6b9-d2f6879175a6","Type":"ContainerDied","Data":"c7a1e801804e99d488b91871797f6061ec5ab018dfd37e42457ab9fe58227712"} Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.698084 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" event={"ID":"6096bf5b-7d9b-41c1-a6b9-d2f6879175a6","Type":"ContainerStarted","Data":"b025eea4bad32b1be23d5b905a056c7f560134ae95b968aca51d37d9e95d4ff8"} Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.698552 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:48 crc kubenswrapper[4934]: E1002 09:50:48.700286 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:49.20027032 +0000 UTC m=+120.952911972 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.723033 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lch8r" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.724554 4934 generic.go:334] "Generic (PLEG): container finished" podID="76e57d57-21b3-4f79-a7ee-b6bf6117c8d3" containerID="ce8ebe92922d677916316a5ced26e9263bf3854f4e9eb3373330d52c1a7ca236" exitCode=0 Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.724868 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vhxn6" event={"ID":"76e57d57-21b3-4f79-a7ee-b6bf6117c8d3","Type":"ContainerDied","Data":"ce8ebe92922d677916316a5ced26e9263bf3854f4e9eb3373330d52c1a7ca236"} Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.724960 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vhxn6" event={"ID":"76e57d57-21b3-4f79-a7ee-b6bf6117c8d3","Type":"ContainerStarted","Data":"6782ca7d9b08a1b97b173fb589b270d5b9865b9a1585333ced5f073bc5f77626"} Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.732471 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-q2p9l" event={"ID":"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9","Type":"ContainerStarted","Data":"6df8a7c62f24bef35b7570cf67499418e438e5978cf6864f33decbe9546a8d3e"} Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.734296 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-q2p9l" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.754673 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z2d6w" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.759372 4934 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-q2p9l container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.759423 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-q2p9l" podUID="f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.765844 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rjhkx"] Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.770780 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-f44r4"] Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.771040 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wlf9l" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.778429 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-w6z57" Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.801359 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:48 crc kubenswrapper[4934]: E1002 09:50:48.801538 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:49.301509706 +0000 UTC m=+121.054151228 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.804261 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:48 crc kubenswrapper[4934]: E1002 09:50:48.804707 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:49.304694529 +0000 UTC m=+121.057336051 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.904940 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:48 crc kubenswrapper[4934]: E1002 09:50:48.905076 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:49.405053696 +0000 UTC m=+121.157695218 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:48 crc kubenswrapper[4934]: I1002 09:50:48.905173 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:48 crc kubenswrapper[4934]: E1002 09:50:48.905462 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:49.405454785 +0000 UTC m=+121.158096307 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:48 crc kubenswrapper[4934]: W1002 09:50:48.928043 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod460de405_cddd_4e59_8be4_537419c25f55.slice/crio-b47a3b2e71d73a999953f47127512541b011bbf4f49f8fcc5dc4f43e0184a24f WatchSource:0}: Error finding container b47a3b2e71d73a999953f47127512541b011bbf4f49f8fcc5dc4f43e0184a24f: Status 404 returned error can't find the container with id b47a3b2e71d73a999953f47127512541b011bbf4f49f8fcc5dc4f43e0184a24f Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.006759 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:49 crc kubenswrapper[4934]: E1002 09:50:49.008461 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:49.508440592 +0000 UTC m=+121.261082114 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.009272 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4pp8p"] Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.015701 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-5wrws"] Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.017224 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-wjz5g" Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.056159 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-66hrr"] Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.107985 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:49 crc kubenswrapper[4934]: E1002 09:50:49.108328 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:49.608312166 +0000 UTC m=+121.360953688 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.209202 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:49 crc kubenswrapper[4934]: E1002 09:50:49.209367 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:49.709340847 +0000 UTC m=+121.461982369 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.209811 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:49 crc kubenswrapper[4934]: E1002 09:50:49.210225 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:49.710208788 +0000 UTC m=+121.462850320 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.311150 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:49 crc kubenswrapper[4934]: E1002 09:50:49.311570 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:49.811550497 +0000 UTC m=+121.564192019 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.413007 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:49 crc kubenswrapper[4934]: E1002 09:50:49.413394 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:49.913379556 +0000 UTC m=+121.666021078 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.534918 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:49 crc kubenswrapper[4934]: E1002 09:50:49.535159 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:50.035127496 +0000 UTC m=+121.787769038 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.550191 4934 patch_prober.go:28] interesting pod/router-default-5444994796-wjz5g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:50:49 crc kubenswrapper[4934]: [-]has-synced failed: reason withheld Oct 02 09:50:49 crc kubenswrapper[4934]: [+]process-running ok Oct 02 09:50:49 crc kubenswrapper[4934]: healthz check failed Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.550241 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wjz5g" podUID="8391ee1a-9c57-4745-825f-381f47254807" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.636949 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:49 crc kubenswrapper[4934]: E1002 09:50:49.637855 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:50.137837726 +0000 UTC m=+121.890479248 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.741677 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:49 crc kubenswrapper[4934]: E1002 09:50:49.742069 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:50.242051781 +0000 UTC m=+121.994693303 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.759947 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7t8rp" event={"ID":"fec65034-065a-466e-8c61-dacc1e415b6f","Type":"ContainerStarted","Data":"77570d641fb9523b3f1fcf92bf24cd6b036383c2ad922130020560b72859c776"} Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.802722 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" event={"ID":"962afee9-b32a-40cc-b6ca-c8e1cb95c13f","Type":"ContainerStarted","Data":"60c397c9dc0b691e29e2158d3fcc21e662f8aa98a09665105e2da4702903a607"} Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.822821 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7qgvq" event={"ID":"fbf54dc7-2324-4259-b987-a0df48e06da5","Type":"ContainerStarted","Data":"6030718e1edef2bb249fe8e8f2bea5d2cc502ee1df789bbca3d66b8064a8c5df"} Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.835582 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-66hrr" event={"ID":"6d137db5-66a8-4066-b8d2-e88adcaaf12c","Type":"ContainerStarted","Data":"3e72008b0d74ac2546100da34321d6fd469d741981a96a52127b165bae10326e"} Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.838739 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxlv9" event={"ID":"da6bd676-392c-4a8a-8a05-dee3304ef594","Type":"ContainerStarted","Data":"f201d103d194c5a3e73b29ea9e02936d94a20960b6c97e014ef740a7c5d915b7"} Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.842188 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb67r" event={"ID":"aa8f023f-df93-4883-9771-b0b23c472246","Type":"ContainerStarted","Data":"cdc5c044afef4ce2c8629766e8b71306f58da7fa33fd4f036c696c61adbdedf9"} Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.847661 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:49 crc kubenswrapper[4934]: E1002 09:50:49.848114 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:50.348098108 +0000 UTC m=+122.100739630 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.851103 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vhxn6" event={"ID":"76e57d57-21b3-4f79-a7ee-b6bf6117c8d3","Type":"ContainerStarted","Data":"134f80c2c6a877da99144ad385f9e257870427a41e1e27633530c2e771d5e9df"} Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.851195 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vhxn6" Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.854981 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bnvc4" event={"ID":"586fee81-a26a-4506-8dfa-66db12798618","Type":"ContainerStarted","Data":"17e56dd28500570bf2a2fbbaa4cb6e7d61cd3639a6b31e03ec2e2eb7239d3f8d"} Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.859986 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rjhkx" event={"ID":"460de405-cddd-4e59-8be4-537419c25f55","Type":"ContainerStarted","Data":"b47a3b2e71d73a999953f47127512541b011bbf4f49f8fcc5dc4f43e0184a24f"} Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.862883 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qg98s" event={"ID":"ebd4ee43-4dc5-46dc-964b-236cd1b16ec4","Type":"ContainerStarted","Data":"3c9ebfad516e84802a52fd0bd4810273f3457d7e328e7e75205d440b3bb8a73f"} Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.862968 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qg98s" event={"ID":"ebd4ee43-4dc5-46dc-964b-236cd1b16ec4","Type":"ContainerStarted","Data":"0ec48d7706c90386a6b09dd3847077f2fd89816a3338ec05e550be8a488f1ef3"} Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.870980 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-5wrws" event={"ID":"5f7c01ef-b9b0-476b-b3e0-12b8132b58b9","Type":"ContainerStarted","Data":"95420c8cc5c3d44cfa276cbf8f0412f49ea5842a75f7ed84c48fadc474e7fd34"} Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.875322 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-w6z57" event={"ID":"9fd0eab0-2348-495b-83cf-580fedb4c1f1","Type":"ContainerStarted","Data":"1119070f5533cda04d208668590efb26738f51ac8274854d7b13daa21e740f3c"} Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.887815 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-f44r4" event={"ID":"ad155aa0-50a3-4874-93bd-2166a8d093cb","Type":"ContainerStarted","Data":"a626b08f6728d9fc8d3ce28ee4ced96065cc97c159063fb5beb5eafd357f65b9"} Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.888782 4934 patch_prober.go:28] interesting pod/downloads-7954f5f757-mcnd2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.888817 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mcnd2" podUID="5dbb4e0e-9898-44ab-81f9-d149b80cfef0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.896645 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-q2p9l" Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.911395 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6" Oct 02 09:50:49 crc kubenswrapper[4934]: I1002 09:50:49.949788 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:49 crc kubenswrapper[4934]: E1002 09:50:49.950948 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:50.450932162 +0000 UTC m=+122.203573684 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.045144 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-t6dmx" podStartSLOduration=100.045101235 podStartE2EDuration="1m40.045101235s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:50.040413207 +0000 UTC m=+121.793054729" watchObservedRunningTime="2025-10-02 09:50:50.045101235 +0000 UTC m=+121.797742767" Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.051260 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:50 crc kubenswrapper[4934]: E1002 09:50:50.051836 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:50.55181966 +0000 UTC m=+122.304461182 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.086212 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-mcnd2" podStartSLOduration=100.086193523 podStartE2EDuration="1m40.086193523s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:50.080073842 +0000 UTC m=+121.832715374" watchObservedRunningTime="2025-10-02 09:50:50.086193523 +0000 UTC m=+121.838835045" Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.145656 4934 patch_prober.go:28] interesting pod/router-default-5444994796-wjz5g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:50:50 crc kubenswrapper[4934]: [-]has-synced failed: reason withheld Oct 02 09:50:50 crc kubenswrapper[4934]: [+]process-running ok Oct 02 09:50:50 crc kubenswrapper[4934]: healthz check failed Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.145712 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wjz5g" podUID="8391ee1a-9c57-4745-825f-381f47254807" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.152555 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:50 crc kubenswrapper[4934]: E1002 09:50:50.166549 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:50.666519597 +0000 UTC m=+122.419161119 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.199548 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gmxzf" podStartSLOduration=101.199525879 podStartE2EDuration="1m41.199525879s" podCreationTimestamp="2025-10-02 09:49:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:50.165828771 +0000 UTC m=+121.918470293" watchObservedRunningTime="2025-10-02 09:50:50.199525879 +0000 UTC m=+121.952167401" Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.200216 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-q2p9l" podStartSLOduration=100.200210855 podStartE2EDuration="1m40.200210855s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:50.195194389 +0000 UTC m=+121.947835911" watchObservedRunningTime="2025-10-02 09:50:50.200210855 +0000 UTC m=+121.952852377" Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.261194 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:50 crc kubenswrapper[4934]: E1002 09:50:50.261522 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:50.761511379 +0000 UTC m=+122.514152901 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.285840 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-wjz5g" podStartSLOduration=100.285815571 podStartE2EDuration="1m40.285815571s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:50.232400627 +0000 UTC m=+121.985042149" watchObservedRunningTime="2025-10-02 09:50:50.285815571 +0000 UTC m=+122.038457093" Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.306186 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-4bqtk" podStartSLOduration=101.30616838 podStartE2EDuration="1m41.30616838s" podCreationTimestamp="2025-10-02 09:49:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:50.306144009 +0000 UTC m=+122.058785531" watchObservedRunningTime="2025-10-02 09:50:50.30616838 +0000 UTC m=+122.058809902" Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.362990 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:50 crc kubenswrapper[4934]: E1002 09:50:50.363251 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:50.863237167 +0000 UTC m=+122.615878689 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.388226 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qg98s" podStartSLOduration=100.388210873 podStartE2EDuration="1m40.388210873s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:50.34694561 +0000 UTC m=+122.099587132" watchObservedRunningTime="2025-10-02 09:50:50.388210873 +0000 UTC m=+122.140852385" Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.389180 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vhxn6" podStartSLOduration=101.389174775 podStartE2EDuration="1m41.389174775s" podCreationTimestamp="2025-10-02 09:49:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:50.387249221 +0000 UTC m=+122.139890743" watchObservedRunningTime="2025-10-02 09:50:50.389174775 +0000 UTC m=+122.141816297" Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.421156 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6" podStartSLOduration=100.421140463 podStartE2EDuration="1m40.421140463s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:50.420885627 +0000 UTC m=+122.173527149" watchObservedRunningTime="2025-10-02 09:50:50.421140463 +0000 UTC m=+122.173781975" Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.464368 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:50 crc kubenswrapper[4934]: E1002 09:50:50.464740 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:50.964727539 +0000 UTC m=+122.717369061 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.567216 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:50 crc kubenswrapper[4934]: E1002 09:50:50.567529 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:51.067501101 +0000 UTC m=+122.820142623 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.567825 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:50 crc kubenswrapper[4934]: E1002 09:50:50.568196 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:51.068181436 +0000 UTC m=+122.820822958 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.668478 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:50 crc kubenswrapper[4934]: E1002 09:50:50.668938 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:51.168915241 +0000 UTC m=+122.921556763 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.770285 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:50 crc kubenswrapper[4934]: E1002 09:50:50.770893 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:51.270881065 +0000 UTC m=+123.023522587 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.873634 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:50 crc kubenswrapper[4934]: E1002 09:50:50.873904 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:51.373890221 +0000 UTC m=+123.126531743 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.893675 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb67r" event={"ID":"aa8f023f-df93-4883-9771-b0b23c472246","Type":"ContainerStarted","Data":"93ed191a051056b02994bd80daa784274b83705a49987f166112a51362742f16"} Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.896471 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-f44r4" event={"ID":"ad155aa0-50a3-4874-93bd-2166a8d093cb","Type":"ContainerStarted","Data":"4a7f374112692e2fb8e8ed98087537f84a98f191b103da6bc4f4be2c2d11d5cc"} Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.902934 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bnvc4" event={"ID":"586fee81-a26a-4506-8dfa-66db12798618","Type":"ContainerStarted","Data":"8a0a33bef6a1f0ed8f7f9e03c5d378b4131bb833dd0b4858e2982eaa23054f6f"} Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.905346 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6zlqv" podStartSLOduration=100.905328227 podStartE2EDuration="1m40.905328227s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:50.511274623 +0000 UTC m=+122.263916145" watchObservedRunningTime="2025-10-02 09:50:50.905328227 +0000 UTC m=+122.657969749" Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.908526 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-86wbg"] Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.923717 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb67r" podStartSLOduration=100.923700901 podStartE2EDuration="1m40.923700901s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:50.919169547 +0000 UTC m=+122.671811089" watchObservedRunningTime="2025-10-02 09:50:50.923700901 +0000 UTC m=+122.676342423" Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.926045 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" event={"ID":"962afee9-b32a-40cc-b6ca-c8e1cb95c13f","Type":"ContainerStarted","Data":"7faa43dce95ff9b51790c77d99013b25fa2bb327e13eae8ad536c5870acd730c"} Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.926408 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.941405 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hw97b"] Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.944715 4934 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-4pp8p container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.34:6443/healthz\": dial tcp 10.217.0.34:6443: connect: connection refused" start-of-body= Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.944774 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" podUID="962afee9-b32a-40cc-b6ca-c8e1cb95c13f" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.34:6443/healthz\": dial tcp 10.217.0.34:6443: connect: connection refused" Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.952723 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7qgvq" event={"ID":"fbf54dc7-2324-4259-b987-a0df48e06da5","Type":"ContainerStarted","Data":"6b747a608276fe879d096a716b8a61e7755f6b6e2fdaf2b47c68736aa1c987ad"} Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.953372 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-7frl9"] Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.956055 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-f44r4" podStartSLOduration=100.956038558 podStartE2EDuration="1m40.956038558s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:50.949627099 +0000 UTC m=+122.702268621" watchObservedRunningTime="2025-10-02 09:50:50.956038558 +0000 UTC m=+122.708680080" Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.977462 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" event={"ID":"6096bf5b-7d9b-41c1-a6b9-d2f6879175a6","Type":"ContainerStarted","Data":"35572cf907d7db76d095a67079af1110b60a33b1dc95a6114b3eeff3cc63d8a6"} Oct 02 09:50:50 crc kubenswrapper[4934]: I1002 09:50:50.979296 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:51 crc kubenswrapper[4934]: E1002 09:50:51.002108 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:51.50207524 +0000 UTC m=+123.254716752 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.006308 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxlv9" event={"ID":"da6bd676-392c-4a8a-8a05-dee3304ef594","Type":"ContainerStarted","Data":"af249d0a2d41bb7c7fcdab7ca7442ebb2383df84547cceb1fbc27da6ca96626a"} Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.029924 4934 patch_prober.go:28] interesting pod/router-default-5444994796-wjz5g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:50:51 crc kubenswrapper[4934]: [-]has-synced failed: reason withheld Oct 02 09:50:51 crc kubenswrapper[4934]: [+]process-running ok Oct 02 09:50:51 crc kubenswrapper[4934]: healthz check failed Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.029981 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wjz5g" podUID="8391ee1a-9c57-4745-825f-381f47254807" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.030537 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" event={"ID":"1d074a40-447d-4144-a2dc-ca69269c84de","Type":"ContainerStarted","Data":"d170dc374e1d44eaa8849462607daae1f2c068a1e493bbc7ade0a5315362ccda"} Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.030570 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" event={"ID":"1d074a40-447d-4144-a2dc-ca69269c84de","Type":"ContainerStarted","Data":"fed858cf3b8ce7e35d14afd4a295ecba0aeb394d91f2ccafee9c7de8f63c47e7"} Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.036477 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-66hrr" event={"ID":"6d137db5-66a8-4066-b8d2-e88adcaaf12c","Type":"ContainerStarted","Data":"cd49a4c22830bb661a44e96674b3557b609b43ff491066ec8fdf67883616f50d"} Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.038803 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-66hrr" Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.058694 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-hr945"] Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.065524 4934 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-66hrr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.066398 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-66hrr" podUID="6d137db5-66a8-4066-b8d2-e88adcaaf12c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.068634 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-pzrnw"] Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.077798 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bnvc4" podStartSLOduration=101.077778667 podStartE2EDuration="1m41.077778667s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:51.039286128 +0000 UTC m=+122.791927650" watchObservedRunningTime="2025-10-02 09:50:51.077778667 +0000 UTC m=+122.830420189" Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.080148 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.082439 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-6jtqh"] Oct 02 09:50:51 crc kubenswrapper[4934]: E1002 09:50:51.082711 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:51.58268694 +0000 UTC m=+123.335328462 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.086272 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rjhkx" event={"ID":"460de405-cddd-4e59-8be4-537419c25f55","Type":"ContainerStarted","Data":"f90e1ea3e740759b8decb4617c87ef94cf565faa590273a59245f4b1958c6976"} Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.088023 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:51 crc kubenswrapper[4934]: E1002 09:50:51.088649 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:51.588626587 +0000 UTC m=+123.341268119 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.093841 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7t8rp" event={"ID":"fec65034-065a-466e-8c61-dacc1e415b6f","Type":"ContainerStarted","Data":"54d102629ca5274200dee8ac35838e22e50ef0cbf1898d41447a3945ceb1453d"} Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.117900 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" podStartSLOduration=102.117873263 podStartE2EDuration="1m42.117873263s" podCreationTimestamp="2025-10-02 09:49:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:51.113482751 +0000 UTC m=+122.866124273" watchObservedRunningTime="2025-10-02 09:50:51.117873263 +0000 UTC m=+122.870514805" Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.117981 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-5wrws" event={"ID":"5f7c01ef-b9b0-476b-b3e0-12b8132b58b9","Type":"ContainerStarted","Data":"ffb779f1f05f6105ca689a6f52c6af7ef2beaff1839ee8a663326ffe01da985e"} Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.118794 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-5wrws" Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.119977 4934 patch_prober.go:28] interesting pod/console-operator-58897d9998-5wrws container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/readyz\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.120019 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-5wrws" podUID="5f7c01ef-b9b0-476b-b3e0-12b8132b58b9" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/readyz\": dial tcp 10.217.0.30:8443: connect: connection refused" Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.125070 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-w6z57" event={"ID":"9fd0eab0-2348-495b-83cf-580fedb4c1f1","Type":"ContainerStarted","Data":"8cc2bea2f05cedbd02ef180ef1acb7f343fb472829eeaea54342de04b5a78130"} Oct 02 09:50:51 crc kubenswrapper[4934]: W1002 09:50:51.159569 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59279e86_a90f_463e_a098_da83a2bdf5a5.slice/crio-9776293fc9e687a4af72123a2a1ff17dfcc7df91e07379a37db5c5c1ef12cdf7 WatchSource:0}: Error finding container 9776293fc9e687a4af72123a2a1ff17dfcc7df91e07379a37db5c5c1ef12cdf7: Status 404 returned error can't find the container with id 9776293fc9e687a4af72123a2a1ff17dfcc7df91e07379a37db5c5c1ef12cdf7 Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.161339 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" podStartSLOduration=101.161317705 podStartE2EDuration="1m41.161317705s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:51.148990461 +0000 UTC m=+122.901631983" watchObservedRunningTime="2025-10-02 09:50:51.161317705 +0000 UTC m=+122.913959237" Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.174035 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-2vn7r"] Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.190518 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:51 crc kubenswrapper[4934]: E1002 09:50:51.192608 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:51.692590046 +0000 UTC m=+123.445231568 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.244429 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" podStartSLOduration=102.244413693 podStartE2EDuration="1m42.244413693s" podCreationTimestamp="2025-10-02 09:49:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:51.208527194 +0000 UTC m=+122.961168726" watchObservedRunningTime="2025-10-02 09:50:51.244413693 +0000 UTC m=+122.997055215" Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.245490 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-66hrr" podStartSLOduration=101.245482167 podStartE2EDuration="1m41.245482167s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:51.24386112 +0000 UTC m=+122.996502642" watchObservedRunningTime="2025-10-02 09:50:51.245482167 +0000 UTC m=+122.998123689" Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.271144 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7qgvq" podStartSLOduration=101.271126079 podStartE2EDuration="1m41.271126079s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:51.268960878 +0000 UTC m=+123.021602410" watchObservedRunningTime="2025-10-02 09:50:51.271126079 +0000 UTC m=+123.023767601" Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.296728 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:51 crc kubenswrapper[4934]: E1002 09:50:51.298335 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:51.798316146 +0000 UTC m=+123.550957668 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.306960 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-x7mrm"] Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.336649 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lch8r"] Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.339052 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pxlv9" podStartSLOduration=102.339030035 podStartE2EDuration="1m42.339030035s" podCreationTimestamp="2025-10-02 09:49:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:51.317503949 +0000 UTC m=+123.070145481" watchObservedRunningTime="2025-10-02 09:50:51.339030035 +0000 UTC m=+123.091671567" Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.372520 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323305-jv64v"] Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.372789 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-w6z57" podStartSLOduration=6.372777225 podStartE2EDuration="6.372777225s" podCreationTimestamp="2025-10-02 09:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:51.348553756 +0000 UTC m=+123.101195278" watchObservedRunningTime="2025-10-02 09:50:51.372777225 +0000 UTC m=+123.125418747" Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.398215 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:51 crc kubenswrapper[4934]: E1002 09:50:51.398629 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:51.898610611 +0000 UTC m=+123.651252143 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.424779 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9ts9g"] Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.432677 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-v58f6"] Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.434941 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-7t8rp" podStartSLOduration=101.434919419 podStartE2EDuration="1m41.434919419s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:51.398090169 +0000 UTC m=+123.150731691" watchObservedRunningTime="2025-10-02 09:50:51.434919419 +0000 UTC m=+123.187560951" Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.452801 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-rjhkx" podStartSLOduration=101.452784111 podStartE2EDuration="1m41.452784111s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:51.451985303 +0000 UTC m=+123.204626835" watchObservedRunningTime="2025-10-02 09:50:51.452784111 +0000 UTC m=+123.205425643" Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.508794 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.510134 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8xjjw"] Oct 02 09:50:51 crc kubenswrapper[4934]: E1002 09:50:51.510489 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:52.010473063 +0000 UTC m=+123.763114585 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.520214 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-llqjv"] Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.522730 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-pfl7n"] Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.529265 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5s688"] Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.536993 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-5wrws" podStartSLOduration=102.536973004 podStartE2EDuration="1m42.536973004s" podCreationTimestamp="2025-10-02 09:49:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:51.517465664 +0000 UTC m=+123.270107186" watchObservedRunningTime="2025-10-02 09:50:51.536973004 +0000 UTC m=+123.289614526" Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.537943 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z2d6w"] Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.567811 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wlf9l"] Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.612811 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:51 crc kubenswrapper[4934]: E1002 09:50:51.613095 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:52.11308027 +0000 UTC m=+123.865721792 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:51 crc kubenswrapper[4934]: W1002 09:50:51.627989 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod327b634c_c056_49ff_a0ad_a178bc67ac48.slice/crio-c606c9fbe6a297f065bec045918a3f48fcb79cfa71881daf7133d0c0e1400d9e WatchSource:0}: Error finding container c606c9fbe6a297f065bec045918a3f48fcb79cfa71881daf7133d0c0e1400d9e: Status 404 returned error can't find the container with id c606c9fbe6a297f065bec045918a3f48fcb79cfa71881daf7133d0c0e1400d9e Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.715660 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:51 crc kubenswrapper[4934]: E1002 09:50:51.715991 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:52.215977015 +0000 UTC m=+123.968618547 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.816626 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:51 crc kubenswrapper[4934]: E1002 09:50:51.816864 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:52.316836333 +0000 UTC m=+124.069477855 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.817159 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:51 crc kubenswrapper[4934]: E1002 09:50:51.817681 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:52.317666651 +0000 UTC m=+124.070308183 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:51 crc kubenswrapper[4934]: I1002 09:50:51.917945 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:51 crc kubenswrapper[4934]: E1002 09:50:51.918299 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:52.418283264 +0000 UTC m=+124.170924776 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.019001 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:52 crc kubenswrapper[4934]: E1002 09:50:52.019603 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:52.519571721 +0000 UTC m=+124.272213243 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.030331 4934 patch_prober.go:28] interesting pod/router-default-5444994796-wjz5g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:50:52 crc kubenswrapper[4934]: [-]has-synced failed: reason withheld Oct 02 09:50:52 crc kubenswrapper[4934]: [+]process-running ok Oct 02 09:50:52 crc kubenswrapper[4934]: healthz check failed Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.030373 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wjz5g" podUID="8391ee1a-9c57-4745-825f-381f47254807" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.119779 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:52 crc kubenswrapper[4934]: E1002 09:50:52.120063 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:52.62004841 +0000 UTC m=+124.372689932 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.147450 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-x7mrm" event={"ID":"ad1a290c-1e04-46fe-8118-92bab2478385","Type":"ContainerStarted","Data":"0d5b73715728a167f001800b96fd4bf76c5254103eb92ee56cf07bdde1e64e28"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.147501 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-x7mrm" event={"ID":"ad1a290c-1e04-46fe-8118-92bab2478385","Type":"ContainerStarted","Data":"a9bae9e90222b4ef2e822a26e3446367aafb09953845c130cd989eaa0ae0a141"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.155222 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z2d6w" event={"ID":"ef34af0b-654c-4519-9292-4e1a833b2ed3","Type":"ContainerStarted","Data":"5913641f58b5c7c10c1367a09351cb44bbd698309babd6fb9a7d3c18208b1e13"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.166645 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.167073 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.174572 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-llqjv" event={"ID":"4b619ff7-c8d0-43ea-9373-18c36b2e160e","Type":"ContainerStarted","Data":"6bad14743534d89ea30a934bb94a75f38217f799308abcbd76914a8de4fbd78c"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.193679 4934 patch_prober.go:28] interesting pod/apiserver-76f77b778f-fc4l5 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.6:8443/livez\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.193743 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" podUID="1d074a40-447d-4144-a2dc-ca69269c84de" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.6:8443/livez\": dial tcp 10.217.0.6:8443: connect: connection refused" Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.209877 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8xjjw" event={"ID":"28aceb4d-e776-4340-9812-8150b2766d5c","Type":"ContainerStarted","Data":"1ee407e6a946cc8105d78fd7f0054794e6cb4d433acc4cd5c57ed157cb1b289e"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.209934 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8xjjw" event={"ID":"28aceb4d-e776-4340-9812-8150b2766d5c","Type":"ContainerStarted","Data":"95047dcd1bd3d905e9ae10b02263393bf02856e4ee9690239a1243722a322818"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.221606 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:52 crc kubenswrapper[4934]: E1002 09:50:52.222029 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:52.722017264 +0000 UTC m=+124.474658786 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.222806 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hw97b" event={"ID":"8549da74-9ee6-46ff-a0dc-5b4af7d6cfa6","Type":"ContainerStarted","Data":"c26ef4431b219668e314cc381a4151b1c577610b24ff498c441a9e3d28943f7e"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.222864 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hw97b" event={"ID":"8549da74-9ee6-46ff-a0dc-5b4af7d6cfa6","Type":"ContainerStarted","Data":"596baeb02954eddb9f91de91ae8228ea4df82963eb5094bdae2808019faab619"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.255147 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hw97b" podStartSLOduration=102.255129398 podStartE2EDuration="1m42.255129398s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:52.253000269 +0000 UTC m=+124.005641791" watchObservedRunningTime="2025-10-02 09:50:52.255129398 +0000 UTC m=+124.007770920" Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.266388 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-pzrnw" event={"ID":"25f4d147-18a5-4d67-99c5-39273eecacc0","Type":"ContainerStarted","Data":"42d0b8ccfb27452533c98fc3eec154030707de2ab880543b1631d5e1ce8592dc"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.266467 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-pzrnw" event={"ID":"25f4d147-18a5-4d67-99c5-39273eecacc0","Type":"ContainerStarted","Data":"e9734c0a3d23322986af9a88df28ec7e817c40a7c29858231dcf0c3a92be3dec"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.291185 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6jtqh" event={"ID":"59279e86-a90f-463e-a098-da83a2bdf5a5","Type":"ContainerStarted","Data":"ecbebff8c956f2ea7f84e41dfcdfe9d98b7b9332ad426782aef16aa0d58c1b6f"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.291506 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6jtqh" event={"ID":"59279e86-a90f-463e-a098-da83a2bdf5a5","Type":"ContainerStarted","Data":"0ad47f9f5b74e67bc1b4b434b6b5af5687aada679b0355dd5f7a40a915c2ce8e"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.291520 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6jtqh" event={"ID":"59279e86-a90f-463e-a098-da83a2bdf5a5","Type":"ContainerStarted","Data":"9776293fc9e687a4af72123a2a1ff17dfcc7df91e07379a37db5c5c1ef12cdf7"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.301305 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-pzrnw" podStartSLOduration=7.301288853 podStartE2EDuration="7.301288853s" podCreationTimestamp="2025-10-02 09:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:52.300486874 +0000 UTC m=+124.053128406" watchObservedRunningTime="2025-10-02 09:50:52.301288853 +0000 UTC m=+124.053930375" Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.314614 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rjhkx" event={"ID":"460de405-cddd-4e59-8be4-537419c25f55","Type":"ContainerStarted","Data":"cd41e43ef86502952bc8da589ec19c49ed7d592d78386a391ad0e9c02557a2ce"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.322473 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:52 crc kubenswrapper[4934]: E1002 09:50:52.322617 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:52.822571564 +0000 UTC m=+124.575213106 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.322806 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:52 crc kubenswrapper[4934]: E1002 09:50:52.323831 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:52.823820143 +0000 UTC m=+124.576461665 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.325339 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lch8r" event={"ID":"88ecaab6-9c13-4813-8698-5b558d32a2e9","Type":"ContainerStarted","Data":"786ed48469ca136c98c7ef4de8330f77ccf2ce532bd85c08ee50a7cc2d545071"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.325372 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lch8r" event={"ID":"88ecaab6-9c13-4813-8698-5b558d32a2e9","Type":"ContainerStarted","Data":"21f051f8df633230ccde88b6964dfb833aeda5591a8c577e614a2b3296170f32"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.327370 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lch8r" Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.328503 4934 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-lch8r container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" start-of-body= Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.328553 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lch8r" podUID="88ecaab6-9c13-4813-8698-5b558d32a2e9" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.336064 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6jtqh" podStartSLOduration=102.335971473 podStartE2EDuration="1m42.335971473s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:52.333896116 +0000 UTC m=+124.086537638" watchObservedRunningTime="2025-10-02 09:50:52.335971473 +0000 UTC m=+124.088612995" Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.358091 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9ts9g" event={"ID":"66183279-91fd-423e-91bd-19b68dda8ef0","Type":"ContainerStarted","Data":"a55ae010bebe46b034c7fbb8ba3c19779d6336cfa0f266a17489cea7c7fcbdf6"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.358151 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9ts9g" Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.359708 4934 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-9ts9g container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.359774 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9ts9g" podUID="66183279-91fd-423e-91bd-19b68dda8ef0" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.378046 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pfl7n" event={"ID":"de99fcca-2d10-4595-8827-410ac05c5f68","Type":"ContainerStarted","Data":"361a0689cfc393e66001d19a5e55211ec222aec29d1da415a35c5007a59a46c3"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.383901 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lch8r" podStartSLOduration=102.383876658 podStartE2EDuration="1m42.383876658s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:52.365920354 +0000 UTC m=+124.118561876" watchObservedRunningTime="2025-10-02 09:50:52.383876658 +0000 UTC m=+124.136518180" Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.385281 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-2vn7r" event={"ID":"8f6c0ab1-dd58-4734-89f8-aefb606078d3","Type":"ContainerStarted","Data":"0b04625cdd4cdec680bf5ffe1a42e70ce936b4b0f3d33a742c34856dfbd1c493"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.385308 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-2vn7r" event={"ID":"8f6c0ab1-dd58-4734-89f8-aefb606078d3","Type":"ContainerStarted","Data":"57970d435a08fb2e0f9780db85d176816dcd797dd4ec5b2792078bcfd69088fd"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.385734 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9ts9g" podStartSLOduration=102.385725151 podStartE2EDuration="1m42.385725151s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:52.384760719 +0000 UTC m=+124.137402241" watchObservedRunningTime="2025-10-02 09:50:52.385725151 +0000 UTC m=+124.138366673" Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.387487 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-hr945" event={"ID":"90d8cb96-af2e-41b8-a405-85cf9e017631","Type":"ContainerStarted","Data":"ea77c11e07bd4194d22d62c45eb524932cb221fe9f3a7bcf6c434116620ba591"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.387509 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-hr945" event={"ID":"90d8cb96-af2e-41b8-a405-85cf9e017631","Type":"ContainerStarted","Data":"acae7b0cc89aa8124dce4af435fc5b54a86f71d2de6afcbc2c9a492f3172adb8"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.394907 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.394944 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.397918 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7frl9" event={"ID":"305740e5-7461-4097-994e-731594c096bb","Type":"ContainerStarted","Data":"37a3e129e5906cb132b6a41aae61db835d596e8296a1be35850b6c46fe341c74"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.397965 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7frl9" event={"ID":"305740e5-7461-4097-994e-731594c096bb","Type":"ContainerStarted","Data":"22bdfcb5c8ac52790de341459fcdf2b9ca5429cb7c7775e609cdcc6206822e9a"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.397974 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7frl9" event={"ID":"305740e5-7461-4097-994e-731594c096bb","Type":"ContainerStarted","Data":"9334abb6b238b8eba8d824e8f6504edeedf3dcb841e56b2403a2d597bd47d850"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.399684 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wlf9l" event={"ID":"327b634c-c056-49ff-a0ad-a178bc67ac48","Type":"ContainerStarted","Data":"c606c9fbe6a297f065bec045918a3f48fcb79cfa71881daf7133d0c0e1400d9e"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.400383 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wlf9l" Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.409446 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5s688" event={"ID":"f4baf828-a5bb-4d7e-a43d-1194d47e4051","Type":"ContainerStarted","Data":"9fecfe6fff2b447f7cebc42c1355c8d211fa19c99e83e1da944f96f5ec86e240"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.409500 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5s688" event={"ID":"f4baf828-a5bb-4d7e-a43d-1194d47e4051","Type":"ContainerStarted","Data":"a7f423b8b1bc8e90fb8068d62955c375805d07a06bbe77426538c50de9fcc83f"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.409822 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5s688" Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.415133 4934 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-wlf9l container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:5443/healthz\": dial tcp 10.217.0.21:5443: connect: connection refused" start-of-body= Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.415364 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wlf9l" podUID="327b634c-c056-49ff-a0ad-a178bc67ac48" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.21:5443/healthz\": dial tcp 10.217.0.21:5443: connect: connection refused" Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.417517 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-2vn7r" podStartSLOduration=102.417507524 podStartE2EDuration="1m42.417507524s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:52.415677173 +0000 UTC m=+124.168318705" watchObservedRunningTime="2025-10-02 09:50:52.417507524 +0000 UTC m=+124.170149046" Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.424121 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:52 crc kubenswrapper[4934]: E1002 09:50:52.425446 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:52.925420057 +0000 UTC m=+124.678061579 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.427907 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-86wbg" event={"ID":"9e12b431-672f-404d-99ba-de5f718f19f2","Type":"ContainerStarted","Data":"78cf2eb549995d1563a8d7ea3da3a3667409fd7a924f4519d83291650c414f56"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.427951 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-86wbg" event={"ID":"9e12b431-672f-404d-99ba-de5f718f19f2","Type":"ContainerStarted","Data":"b48dc9cfe96e6a59a69a7d8dbe88dd5308fbd31a3fb694e8b051661241168865"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.438059 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-jv64v" event={"ID":"c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a","Type":"ContainerStarted","Data":"0de8328c21c16ce1160c7e885fa77ebef1fa96de06d43610aad59115c78bf81e"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.438086 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-jv64v" event={"ID":"c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a","Type":"ContainerStarted","Data":"b57e8d69babf273a49c2f378fa8455b2d149d4697bca4d2369bac71a0c97d6e5"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.439795 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-v58f6" event={"ID":"dcb231bb-d375-49d3-b967-dc739da957fb","Type":"ContainerStarted","Data":"90bf5774f73060f8bb7b9041158a8bee0190ba5e1ea0f645559efdd38925e221"} Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.444818 4934 patch_prober.go:28] interesting pod/console-operator-58897d9998-5wrws container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/readyz\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.444882 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-5wrws" podUID="5f7c01ef-b9b0-476b-b3e0-12b8132b58b9" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/readyz\": dial tcp 10.217.0.30:8443: connect: connection refused" Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.444954 4934 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-4pp8p container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.34:6443/healthz\": dial tcp 10.217.0.34:6443: connect: connection refused" start-of-body= Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.444982 4934 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-66hrr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.445000 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-66hrr" podUID="6d137db5-66a8-4066-b8d2-e88adcaaf12c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.445009 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" podUID="962afee9-b32a-40cc-b6ca-c8e1cb95c13f" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.34:6443/healthz\": dial tcp 10.217.0.34:6443: connect: connection refused" Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.479042 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5s688" podStartSLOduration=102.479027564 podStartE2EDuration="1m42.479027564s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:52.477691634 +0000 UTC m=+124.230333156" watchObservedRunningTime="2025-10-02 09:50:52.479027564 +0000 UTC m=+124.231669086" Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.535060 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:52 crc kubenswrapper[4934]: E1002 09:50:52.541345 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:53.041330833 +0000 UTC m=+124.793972355 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.602475 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-7frl9" podStartSLOduration=102.602454332 podStartE2EDuration="1m42.602454332s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:52.535882537 +0000 UTC m=+124.288524059" watchObservedRunningTime="2025-10-02 09:50:52.602454332 +0000 UTC m=+124.355095864" Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.602611 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wlf9l" podStartSLOduration=102.602606487 podStartE2EDuration="1m42.602606487s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:52.585262176 +0000 UTC m=+124.337903698" watchObservedRunningTime="2025-10-02 09:50:52.602606487 +0000 UTC m=+124.355248009" Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.641319 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:52 crc kubenswrapper[4934]: E1002 09:50:52.641692 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:53.141676548 +0000 UTC m=+124.894318070 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.660518 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-86wbg" podStartSLOduration=102.660499412 podStartE2EDuration="1m42.660499412s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:52.657945133 +0000 UTC m=+124.410586655" watchObservedRunningTime="2025-10-02 09:50:52.660499412 +0000 UTC m=+124.413140934" Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.743314 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:52 crc kubenswrapper[4934]: E1002 09:50:52.743718 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:53.243703083 +0000 UTC m=+124.996344615 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.844810 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:52 crc kubenswrapper[4934]: E1002 09:50:52.845000 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:53.34497513 +0000 UTC m=+125.097616652 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.845367 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:52 crc kubenswrapper[4934]: E1002 09:50:52.845709 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:53.345698716 +0000 UTC m=+125.098340238 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.862046 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.892146 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-jv64v" podStartSLOduration=102.892126448 podStartE2EDuration="1m42.892126448s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:52.67902127 +0000 UTC m=+124.431662792" watchObservedRunningTime="2025-10-02 09:50:52.892126448 +0000 UTC m=+124.644767970" Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.948156 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:52 crc kubenswrapper[4934]: E1002 09:50:52.948368 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:53.448339295 +0000 UTC m=+125.200980817 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:52 crc kubenswrapper[4934]: I1002 09:50:52.948547 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:52 crc kubenswrapper[4934]: E1002 09:50:52.948980 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:53.448959889 +0000 UTC m=+125.201601481 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.020435 4934 patch_prober.go:28] interesting pod/router-default-5444994796-wjz5g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:50:53 crc kubenswrapper[4934]: [-]has-synced failed: reason withheld Oct 02 09:50:53 crc kubenswrapper[4934]: [+]process-running ok Oct 02 09:50:53 crc kubenswrapper[4934]: healthz check failed Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.020528 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wjz5g" podUID="8391ee1a-9c57-4745-825f-381f47254807" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.050141 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:53 crc kubenswrapper[4934]: E1002 09:50:53.050356 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:53.550329348 +0000 UTC m=+125.302970870 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.050641 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:53 crc kubenswrapper[4934]: E1002 09:50:53.050972 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:53.550957814 +0000 UTC m=+125.303599326 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.151532 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:53 crc kubenswrapper[4934]: E1002 09:50:53.151674 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:53.651656548 +0000 UTC m=+125.404298070 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.151797 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:53 crc kubenswrapper[4934]: E1002 09:50:53.152103 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:53.652095108 +0000 UTC m=+125.404736620 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.252882 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:53 crc kubenswrapper[4934]: E1002 09:50:53.253039 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:53.753018017 +0000 UTC m=+125.505659539 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.253088 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:53 crc kubenswrapper[4934]: E1002 09:50:53.253409 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:53.753399315 +0000 UTC m=+125.506040847 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.326327 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vhxn6" Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.354162 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:53 crc kubenswrapper[4934]: E1002 09:50:53.354387 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:53.854355245 +0000 UTC m=+125.606996767 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.354783 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:53 crc kubenswrapper[4934]: E1002 09:50:53.355137 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:53.855127743 +0000 UTC m=+125.607769265 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.447511 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5s688" event={"ID":"f4baf828-a5bb-4d7e-a43d-1194d47e4051","Type":"ContainerStarted","Data":"86641bb92e7e00203fb447ed20485cd362faa6e4453b73c30f1ded93c2c60b8f"} Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.449620 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8xjjw" event={"ID":"28aceb4d-e776-4340-9812-8150b2766d5c","Type":"ContainerStarted","Data":"5fd525dd81b73e883f13f8da26db13c9cd6683852f3a0928981fa5831a33fe3b"} Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.451912 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-v58f6" event={"ID":"dcb231bb-d375-49d3-b967-dc739da957fb","Type":"ContainerStarted","Data":"f16e108cd9187ba4fc96e6cd3fbbe61d15ced2bdff33f38a627724120073821a"} Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.451955 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-v58f6" event={"ID":"dcb231bb-d375-49d3-b967-dc739da957fb","Type":"ContainerStarted","Data":"f3bbbbdc2c21ee420bf9e0560fd6de134e85253f05b6ff72174980039242212b"} Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.451973 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-v58f6" Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.453402 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-x7mrm" event={"ID":"ad1a290c-1e04-46fe-8118-92bab2478385","Type":"ContainerStarted","Data":"522a760e1fdd36c9144527d6a131485dcc9c230cad09decebfea903f546b23b8"} Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.456254 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:53 crc kubenswrapper[4934]: E1002 09:50:53.456398 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:53.956374379 +0000 UTC m=+125.709015911 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.456556 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.456917 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pfl7n" event={"ID":"de99fcca-2d10-4595-8827-410ac05c5f68","Type":"ContainerStarted","Data":"c082ef1452936dfd384a06d6b13400c61ec3a6f5cebd73dda8386b72d7123e16"} Oct 02 09:50:53 crc kubenswrapper[4934]: E1002 09:50:53.457039 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:53.957027775 +0000 UTC m=+125.709669327 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.458340 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wlf9l" event={"ID":"327b634c-c056-49ff-a0ad-a178bc67ac48","Type":"ContainerStarted","Data":"a1764332cc5ecb17bf8a61065c67cfb1863f6d24e5c56ba36a23e45c454e4d1c"} Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.458900 4934 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-wlf9l container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:5443/healthz\": dial tcp 10.217.0.21:5443: connect: connection refused" start-of-body= Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.458950 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wlf9l" podUID="327b634c-c056-49ff-a0ad-a178bc67ac48" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.21:5443/healthz\": dial tcp 10.217.0.21:5443: connect: connection refused" Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.464089 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8xjjw" podStartSLOduration=103.464073417 podStartE2EDuration="1m43.464073417s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:53.46162079 +0000 UTC m=+125.214262322" watchObservedRunningTime="2025-10-02 09:50:53.464073417 +0000 UTC m=+125.216714939" Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.471143 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z2d6w" event={"ID":"ef34af0b-654c-4519-9292-4e1a833b2ed3","Type":"ContainerStarted","Data":"853b5b329ee640069f8e9b3db49c351871309223f331a88df5d028cb2ad0c243"} Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.475105 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-llqjv" event={"ID":"4b619ff7-c8d0-43ea-9373-18c36b2e160e","Type":"ContainerStarted","Data":"1930c4c4e45bb27cb4485d540ae3d0dcb9da788c478b610480007d94eb7c5125"} Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.477525 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-hr945" event={"ID":"90d8cb96-af2e-41b8-a405-85cf9e017631","Type":"ContainerStarted","Data":"d8cef2e14882edb8f8fd6c1d81e8e80a2c29d84659039a47639a3046201e4c40"} Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.481109 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9ts9g" event={"ID":"66183279-91fd-423e-91bd-19b68dda8ef0","Type":"ContainerStarted","Data":"d2d727b0614a2eeda22d8e02681adf84213a42e2266944f1dee33d37ed16f58e"} Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.481359 4934 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-9ts9g container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.481424 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9ts9g" podUID="66183279-91fd-423e-91bd-19b68dda8ef0" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.483192 4934 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-lch8r container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" start-of-body= Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.483238 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lch8r" podUID="88ecaab6-9c13-4813-8698-5b558d32a2e9" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.497604 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lfhf5" Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.523106 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-v58f6" podStartSLOduration=8.523083869 podStartE2EDuration="8.523083869s" podCreationTimestamp="2025-10-02 09:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:53.518139234 +0000 UTC m=+125.270780766" watchObservedRunningTime="2025-10-02 09:50:53.523083869 +0000 UTC m=+125.275725391" Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.551648 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-x7mrm" podStartSLOduration=103.551632647 podStartE2EDuration="1m43.551632647s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:53.544295869 +0000 UTC m=+125.296937391" watchObservedRunningTime="2025-10-02 09:50:53.551632647 +0000 UTC m=+125.304274169" Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.558278 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:53 crc kubenswrapper[4934]: E1002 09:50:53.560138 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:54.060116813 +0000 UTC m=+125.812758335 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.639992 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-5wrws" Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.642915 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-hr945" podStartSLOduration=103.642892624 podStartE2EDuration="1m43.642892624s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:53.577995136 +0000 UTC m=+125.330636648" watchObservedRunningTime="2025-10-02 09:50:53.642892624 +0000 UTC m=+125.395534146" Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.663361 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:53 crc kubenswrapper[4934]: E1002 09:50:53.665221 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:54.165206499 +0000 UTC m=+125.917848021 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.678111 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z2d6w" podStartSLOduration=103.678088286 podStartE2EDuration="1m43.678088286s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:53.675073216 +0000 UTC m=+125.427714748" watchObservedRunningTime="2025-10-02 09:50:53.678088286 +0000 UTC m=+125.430729818" Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.678377 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-llqjv" podStartSLOduration=103.678369733 podStartE2EDuration="1m43.678369733s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:53.643540039 +0000 UTC m=+125.396181561" watchObservedRunningTime="2025-10-02 09:50:53.678369733 +0000 UTC m=+125.431011255" Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.769049 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:53 crc kubenswrapper[4934]: E1002 09:50:53.769677 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:54.269661789 +0000 UTC m=+126.022303311 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.769715 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:53 crc kubenswrapper[4934]: E1002 09:50:53.769978 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:54.269970786 +0000 UTC m=+126.022612308 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.870851 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:53 crc kubenswrapper[4934]: E1002 09:50:53.871045 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:54.371016178 +0000 UTC m=+126.123657700 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.871194 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:53 crc kubenswrapper[4934]: E1002 09:50:53.871529 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:54.371520831 +0000 UTC m=+126.124162353 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.972680 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:53 crc kubenswrapper[4934]: E1002 09:50:53.972893 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:54.472862019 +0000 UTC m=+126.225503551 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:53 crc kubenswrapper[4934]: I1002 09:50:53.973090 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:53 crc kubenswrapper[4934]: E1002 09:50:53.973463 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:54.473448443 +0000 UTC m=+126.226090025 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.019793 4934 patch_prober.go:28] interesting pod/router-default-5444994796-wjz5g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:50:54 crc kubenswrapper[4934]: [-]has-synced failed: reason withheld Oct 02 09:50:54 crc kubenswrapper[4934]: [+]process-running ok Oct 02 09:50:54 crc kubenswrapper[4934]: healthz check failed Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.019870 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wjz5g" podUID="8391ee1a-9c57-4745-825f-381f47254807" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.073750 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:54 crc kubenswrapper[4934]: E1002 09:50:54.073934 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:54.573901161 +0000 UTC m=+126.326542693 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.074054 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:54 crc kubenswrapper[4934]: E1002 09:50:54.074338 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:54.57432703 +0000 UTC m=+126.326968552 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.174618 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:54 crc kubenswrapper[4934]: E1002 09:50:54.175053 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:54.675027684 +0000 UTC m=+126.427669206 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.276539 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:54 crc kubenswrapper[4934]: E1002 09:50:54.276957 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:54.776937906 +0000 UTC m=+126.529579428 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.365803 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.366742 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.372611 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.372721 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.377575 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:54 crc kubenswrapper[4934]: E1002 09:50:54.377699 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:54.877673871 +0000 UTC m=+126.630315383 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.378243 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:54 crc kubenswrapper[4934]: E1002 09:50:54.378683 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:54.878666814 +0000 UTC m=+126.631308346 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.388923 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.479634 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.479840 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb154e95-c062-4838-ba59-288e60c9138d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"eb154e95-c062-4838-ba59-288e60c9138d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:50:54 crc kubenswrapper[4934]: E1002 09:50:54.479872 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:54.979841669 +0000 UTC m=+126.732483191 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.480063 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.480174 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/eb154e95-c062-4838-ba59-288e60c9138d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"eb154e95-c062-4838-ba59-288e60c9138d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:50:54 crc kubenswrapper[4934]: E1002 09:50:54.480463 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:54.980446093 +0000 UTC m=+126.733087615 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.497095 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lch8r" Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.513818 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9ts9g" Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.581831 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.582257 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/eb154e95-c062-4838-ba59-288e60c9138d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"eb154e95-c062-4838-ba59-288e60c9138d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.582472 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb154e95-c062-4838-ba59-288e60c9138d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"eb154e95-c062-4838-ba59-288e60c9138d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:50:54 crc kubenswrapper[4934]: E1002 09:50:54.583221 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:55.083200604 +0000 UTC m=+126.835842126 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.585892 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/eb154e95-c062-4838-ba59-288e60c9138d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"eb154e95-c062-4838-ba59-288e60c9138d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.647492 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb154e95-c062-4838-ba59-288e60c9138d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"eb154e95-c062-4838-ba59-288e60c9138d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.683962 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.684547 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:54 crc kubenswrapper[4934]: E1002 09:50:54.684878 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:55.18486366 +0000 UTC m=+126.937505182 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.785964 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:54 crc kubenswrapper[4934]: E1002 09:50:54.786258 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:55.28624347 +0000 UTC m=+127.038884992 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.894128 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:54 crc kubenswrapper[4934]: E1002 09:50:54.894818 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:55.394803115 +0000 UTC m=+127.147444637 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:54 crc kubenswrapper[4934]: I1002 09:50:54.995826 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:54 crc kubenswrapper[4934]: E1002 09:50:54.996251 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:55.496234966 +0000 UTC m=+127.248876478 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:55 crc kubenswrapper[4934]: I1002 09:50:55.023807 4934 patch_prober.go:28] interesting pod/router-default-5444994796-wjz5g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:50:55 crc kubenswrapper[4934]: [-]has-synced failed: reason withheld Oct 02 09:50:55 crc kubenswrapper[4934]: [+]process-running ok Oct 02 09:50:55 crc kubenswrapper[4934]: healthz check failed Oct 02 09:50:55 crc kubenswrapper[4934]: I1002 09:50:55.023869 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wjz5g" podUID="8391ee1a-9c57-4745-825f-381f47254807" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:50:55 crc kubenswrapper[4934]: I1002 09:50:55.074935 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wlf9l" Oct 02 09:50:55 crc kubenswrapper[4934]: I1002 09:50:55.097166 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:55 crc kubenswrapper[4934]: E1002 09:50:55.097489 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:55.597474482 +0000 UTC m=+127.350116004 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:55 crc kubenswrapper[4934]: I1002 09:50:55.131308 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 09:50:55 crc kubenswrapper[4934]: W1002 09:50:55.155931 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podeb154e95_c062_4838_ba59_288e60c9138d.slice/crio-25fb10aaba38cb2c0c5d4bd4ec40b70b80e2ec9a91f54c9567afbb14fd1fc743 WatchSource:0}: Error finding container 25fb10aaba38cb2c0c5d4bd4ec40b70b80e2ec9a91f54c9567afbb14fd1fc743: Status 404 returned error can't find the container with id 25fb10aaba38cb2c0c5d4bd4ec40b70b80e2ec9a91f54c9567afbb14fd1fc743 Oct 02 09:50:55 crc kubenswrapper[4934]: I1002 09:50:55.204122 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:55 crc kubenswrapper[4934]: E1002 09:50:55.204509 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:55.704490301 +0000 UTC m=+127.457131823 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:55 crc kubenswrapper[4934]: I1002 09:50:55.297326 4934 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 02 09:50:55 crc kubenswrapper[4934]: I1002 09:50:55.305546 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:55 crc kubenswrapper[4934]: E1002 09:50:55.305836 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:55.805826091 +0000 UTC m=+127.558467613 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:55 crc kubenswrapper[4934]: I1002 09:50:55.406370 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:55 crc kubenswrapper[4934]: E1002 09:50:55.406729 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:55.906714439 +0000 UTC m=+127.659355951 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:55 crc kubenswrapper[4934]: I1002 09:50:55.491448 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"eb154e95-c062-4838-ba59-288e60c9138d","Type":"ContainerStarted","Data":"25fb10aaba38cb2c0c5d4bd4ec40b70b80e2ec9a91f54c9567afbb14fd1fc743"} Oct 02 09:50:55 crc kubenswrapper[4934]: I1002 09:50:55.494569 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pfl7n" event={"ID":"de99fcca-2d10-4595-8827-410ac05c5f68","Type":"ContainerStarted","Data":"962d49d9f41d89ffb848362b8d5e1ad675b4a09df97205af7b39241b9df6e87e"} Oct 02 09:50:55 crc kubenswrapper[4934]: I1002 09:50:55.494640 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pfl7n" event={"ID":"de99fcca-2d10-4595-8827-410ac05c5f68","Type":"ContainerStarted","Data":"3ef13a84dfe36bd5259837d06a1d05489d7a3dbc07affa150073ea416f5f31e6"} Oct 02 09:50:55 crc kubenswrapper[4934]: I1002 09:50:55.508922 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:55 crc kubenswrapper[4934]: E1002 09:50:55.509257 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:56.009241395 +0000 UTC m=+127.761882917 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:55 crc kubenswrapper[4934]: I1002 09:50:55.609892 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:55 crc kubenswrapper[4934]: E1002 09:50:55.611430 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:56.111414113 +0000 UTC m=+127.864055635 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:55 crc kubenswrapper[4934]: I1002 09:50:55.711340 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:55 crc kubenswrapper[4934]: E1002 09:50:55.711766 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:56.211751938 +0000 UTC m=+127.964393470 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:55 crc kubenswrapper[4934]: I1002 09:50:55.812831 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:55 crc kubenswrapper[4934]: E1002 09:50:55.813248 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:56.313227611 +0000 UTC m=+128.065869133 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:55 crc kubenswrapper[4934]: I1002 09:50:55.914197 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:55 crc kubenswrapper[4934]: E1002 09:50:55.914510 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:56.414498858 +0000 UTC m=+128.167140380 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:55 crc kubenswrapper[4934]: I1002 09:50:55.953832 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6k6dn"] Oct 02 09:50:55 crc kubenswrapper[4934]: I1002 09:50:55.955081 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6k6dn" Oct 02 09:50:55 crc kubenswrapper[4934]: I1002 09:50:55.957035 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 02 09:50:55 crc kubenswrapper[4934]: I1002 09:50:55.969810 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6k6dn"] Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.015736 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:56 crc kubenswrapper[4934]: E1002 09:50:56.016226 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:50:56.516206214 +0000 UTC m=+128.268847746 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.020617 4934 patch_prober.go:28] interesting pod/router-default-5444994796-wjz5g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:50:56 crc kubenswrapper[4934]: [-]has-synced failed: reason withheld Oct 02 09:50:56 crc kubenswrapper[4934]: [+]process-running ok Oct 02 09:50:56 crc kubenswrapper[4934]: healthz check failed Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.020669 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wjz5g" podUID="8391ee1a-9c57-4745-825f-381f47254807" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.118035 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea34a4e8-c842-4a23-933a-8ecd3f882d19-utilities\") pod \"community-operators-6k6dn\" (UID: \"ea34a4e8-c842-4a23-933a-8ecd3f882d19\") " pod="openshift-marketplace/community-operators-6k6dn" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.118640 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea34a4e8-c842-4a23-933a-8ecd3f882d19-catalog-content\") pod \"community-operators-6k6dn\" (UID: \"ea34a4e8-c842-4a23-933a-8ecd3f882d19\") " pod="openshift-marketplace/community-operators-6k6dn" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.118769 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqmfs\" (UniqueName: \"kubernetes.io/projected/ea34a4e8-c842-4a23-933a-8ecd3f882d19-kube-api-access-vqmfs\") pod \"community-operators-6k6dn\" (UID: \"ea34a4e8-c842-4a23-933a-8ecd3f882d19\") " pod="openshift-marketplace/community-operators-6k6dn" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.118821 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:56 crc kubenswrapper[4934]: E1002 09:50:56.119244 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:50:56.619223792 +0000 UTC m=+128.371865314 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-jht5p" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.124007 4934 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-02T09:50:55.29756564Z","Handler":null,"Name":""} Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.129644 4934 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.129693 4934 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.154847 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fdrpb"] Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.155860 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fdrpb" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.159178 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.169938 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fdrpb"] Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.224366 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.224827 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea34a4e8-c842-4a23-933a-8ecd3f882d19-utilities\") pod \"community-operators-6k6dn\" (UID: \"ea34a4e8-c842-4a23-933a-8ecd3f882d19\") " pod="openshift-marketplace/community-operators-6k6dn" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.224878 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea34a4e8-c842-4a23-933a-8ecd3f882d19-catalog-content\") pod \"community-operators-6k6dn\" (UID: \"ea34a4e8-c842-4a23-933a-8ecd3f882d19\") " pod="openshift-marketplace/community-operators-6k6dn" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.225001 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqmfs\" (UniqueName: \"kubernetes.io/projected/ea34a4e8-c842-4a23-933a-8ecd3f882d19-kube-api-access-vqmfs\") pod \"community-operators-6k6dn\" (UID: \"ea34a4e8-c842-4a23-933a-8ecd3f882d19\") " pod="openshift-marketplace/community-operators-6k6dn" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.226056 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea34a4e8-c842-4a23-933a-8ecd3f882d19-utilities\") pod \"community-operators-6k6dn\" (UID: \"ea34a4e8-c842-4a23-933a-8ecd3f882d19\") " pod="openshift-marketplace/community-operators-6k6dn" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.226356 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea34a4e8-c842-4a23-933a-8ecd3f882d19-catalog-content\") pod \"community-operators-6k6dn\" (UID: \"ea34a4e8-c842-4a23-933a-8ecd3f882d19\") " pod="openshift-marketplace/community-operators-6k6dn" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.238197 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.254052 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqmfs\" (UniqueName: \"kubernetes.io/projected/ea34a4e8-c842-4a23-933a-8ecd3f882d19-kube-api-access-vqmfs\") pod \"community-operators-6k6dn\" (UID: \"ea34a4e8-c842-4a23-933a-8ecd3f882d19\") " pod="openshift-marketplace/community-operators-6k6dn" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.270863 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6k6dn" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.326121 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8150488-b2c9-45e0-8209-9d6a73b64834-utilities\") pod \"certified-operators-fdrpb\" (UID: \"b8150488-b2c9-45e0-8209-9d6a73b64834\") " pod="openshift-marketplace/certified-operators-fdrpb" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.326237 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8150488-b2c9-45e0-8209-9d6a73b64834-catalog-content\") pod \"certified-operators-fdrpb\" (UID: \"b8150488-b2c9-45e0-8209-9d6a73b64834\") " pod="openshift-marketplace/certified-operators-fdrpb" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.326299 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svl4c\" (UniqueName: \"kubernetes.io/projected/b8150488-b2c9-45e0-8209-9d6a73b64834-kube-api-access-svl4c\") pod \"certified-operators-fdrpb\" (UID: \"b8150488-b2c9-45e0-8209-9d6a73b64834\") " pod="openshift-marketplace/certified-operators-fdrpb" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.326368 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.328864 4934 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.328905 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.364721 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-jht5p\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.368104 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fj8zn"] Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.369218 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fj8zn" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.370916 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fj8zn"] Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.389799 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.398836 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.433033 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8150488-b2c9-45e0-8209-9d6a73b64834-catalog-content\") pod \"certified-operators-fdrpb\" (UID: \"b8150488-b2c9-45e0-8209-9d6a73b64834\") " pod="openshift-marketplace/certified-operators-fdrpb" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.433420 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svl4c\" (UniqueName: \"kubernetes.io/projected/b8150488-b2c9-45e0-8209-9d6a73b64834-kube-api-access-svl4c\") pod \"certified-operators-fdrpb\" (UID: \"b8150488-b2c9-45e0-8209-9d6a73b64834\") " pod="openshift-marketplace/certified-operators-fdrpb" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.433479 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8150488-b2c9-45e0-8209-9d6a73b64834-utilities\") pod \"certified-operators-fdrpb\" (UID: \"b8150488-b2c9-45e0-8209-9d6a73b64834\") " pod="openshift-marketplace/certified-operators-fdrpb" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.434173 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8150488-b2c9-45e0-8209-9d6a73b64834-utilities\") pod \"certified-operators-fdrpb\" (UID: \"b8150488-b2c9-45e0-8209-9d6a73b64834\") " pod="openshift-marketplace/certified-operators-fdrpb" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.434390 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8150488-b2c9-45e0-8209-9d6a73b64834-catalog-content\") pod \"certified-operators-fdrpb\" (UID: \"b8150488-b2c9-45e0-8209-9d6a73b64834\") " pod="openshift-marketplace/certified-operators-fdrpb" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.453384 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svl4c\" (UniqueName: \"kubernetes.io/projected/b8150488-b2c9-45e0-8209-9d6a73b64834-kube-api-access-svl4c\") pod \"certified-operators-fdrpb\" (UID: \"b8150488-b2c9-45e0-8209-9d6a73b64834\") " pod="openshift-marketplace/certified-operators-fdrpb" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.470450 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fdrpb" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.508130 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6k6dn"] Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.511640 4934 generic.go:334] "Generic (PLEG): container finished" podID="eb154e95-c062-4838-ba59-288e60c9138d" containerID="e770280541920af0d61ef6e10c08e4e6f35c19837d512430c4506bcd9f4ff8cf" exitCode=0 Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.511738 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"eb154e95-c062-4838-ba59-288e60c9138d","Type":"ContainerDied","Data":"e770280541920af0d61ef6e10c08e4e6f35c19837d512430c4506bcd9f4ff8cf"} Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.535368 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab1258e2-b9fc-44da-8e7d-c9e70cd7897d-catalog-content\") pod \"community-operators-fj8zn\" (UID: \"ab1258e2-b9fc-44da-8e7d-c9e70cd7897d\") " pod="openshift-marketplace/community-operators-fj8zn" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.535449 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab1258e2-b9fc-44da-8e7d-c9e70cd7897d-utilities\") pod \"community-operators-fj8zn\" (UID: \"ab1258e2-b9fc-44da-8e7d-c9e70cd7897d\") " pod="openshift-marketplace/community-operators-fj8zn" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.535483 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj2fk\" (UniqueName: \"kubernetes.io/projected/ab1258e2-b9fc-44da-8e7d-c9e70cd7897d-kube-api-access-wj2fk\") pod \"community-operators-fj8zn\" (UID: \"ab1258e2-b9fc-44da-8e7d-c9e70cd7897d\") " pod="openshift-marketplace/community-operators-fj8zn" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.558092 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pfl7n" event={"ID":"de99fcca-2d10-4595-8827-410ac05c5f68","Type":"ContainerStarted","Data":"3ba53d22faec191bd0c31b94dfd6b500d58b0a121b6738b42c034f648148d98f"} Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.560439 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-llmzg"] Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.561674 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-llmzg" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.572840 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-llmzg"] Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.578685 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-pfl7n" podStartSLOduration=11.578669945 podStartE2EDuration="11.578669945s" podCreationTimestamp="2025-10-02 09:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:56.575943682 +0000 UTC m=+128.328585214" watchObservedRunningTime="2025-10-02 09:50:56.578669945 +0000 UTC m=+128.331311467" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.638153 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab1258e2-b9fc-44da-8e7d-c9e70cd7897d-catalog-content\") pod \"community-operators-fj8zn\" (UID: \"ab1258e2-b9fc-44da-8e7d-c9e70cd7897d\") " pod="openshift-marketplace/community-operators-fj8zn" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.638509 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab1258e2-b9fc-44da-8e7d-c9e70cd7897d-utilities\") pod \"community-operators-fj8zn\" (UID: \"ab1258e2-b9fc-44da-8e7d-c9e70cd7897d\") " pod="openshift-marketplace/community-operators-fj8zn" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.638536 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj2fk\" (UniqueName: \"kubernetes.io/projected/ab1258e2-b9fc-44da-8e7d-c9e70cd7897d-kube-api-access-wj2fk\") pod \"community-operators-fj8zn\" (UID: \"ab1258e2-b9fc-44da-8e7d-c9e70cd7897d\") " pod="openshift-marketplace/community-operators-fj8zn" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.638730 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab1258e2-b9fc-44da-8e7d-c9e70cd7897d-catalog-content\") pod \"community-operators-fj8zn\" (UID: \"ab1258e2-b9fc-44da-8e7d-c9e70cd7897d\") " pod="openshift-marketplace/community-operators-fj8zn" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.638956 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab1258e2-b9fc-44da-8e7d-c9e70cd7897d-utilities\") pod \"community-operators-fj8zn\" (UID: \"ab1258e2-b9fc-44da-8e7d-c9e70cd7897d\") " pod="openshift-marketplace/community-operators-fj8zn" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.664909 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj2fk\" (UniqueName: \"kubernetes.io/projected/ab1258e2-b9fc-44da-8e7d-c9e70cd7897d-kube-api-access-wj2fk\") pod \"community-operators-fj8zn\" (UID: \"ab1258e2-b9fc-44da-8e7d-c9e70cd7897d\") " pod="openshift-marketplace/community-operators-fj8zn" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.665006 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jht5p"] Oct 02 09:50:56 crc kubenswrapper[4934]: W1002 09:50:56.688011 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod39e0228e_f4f6_4167_9bbc_d319374345a1.slice/crio-9c6741c48da97be054129169423fe482a7880193c8c331dd0144ab1ae4dba703 WatchSource:0}: Error finding container 9c6741c48da97be054129169423fe482a7880193c8c331dd0144ab1ae4dba703: Status 404 returned error can't find the container with id 9c6741c48da97be054129169423fe482a7880193c8c331dd0144ab1ae4dba703 Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.690155 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fj8zn" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.739427 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0220445-630a-4d62-91c9-ba260e6ba945-catalog-content\") pod \"certified-operators-llmzg\" (UID: \"e0220445-630a-4d62-91c9-ba260e6ba945\") " pod="openshift-marketplace/certified-operators-llmzg" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.739547 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0220445-630a-4d62-91c9-ba260e6ba945-utilities\") pod \"certified-operators-llmzg\" (UID: \"e0220445-630a-4d62-91c9-ba260e6ba945\") " pod="openshift-marketplace/certified-operators-llmzg" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.739588 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jsjc\" (UniqueName: \"kubernetes.io/projected/e0220445-630a-4d62-91c9-ba260e6ba945-kube-api-access-2jsjc\") pod \"certified-operators-llmzg\" (UID: \"e0220445-630a-4d62-91c9-ba260e6ba945\") " pod="openshift-marketplace/certified-operators-llmzg" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.834958 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fdrpb"] Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.840433 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jsjc\" (UniqueName: \"kubernetes.io/projected/e0220445-630a-4d62-91c9-ba260e6ba945-kube-api-access-2jsjc\") pod \"certified-operators-llmzg\" (UID: \"e0220445-630a-4d62-91c9-ba260e6ba945\") " pod="openshift-marketplace/certified-operators-llmzg" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.840495 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0220445-630a-4d62-91c9-ba260e6ba945-catalog-content\") pod \"certified-operators-llmzg\" (UID: \"e0220445-630a-4d62-91c9-ba260e6ba945\") " pod="openshift-marketplace/certified-operators-llmzg" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.840611 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0220445-630a-4d62-91c9-ba260e6ba945-utilities\") pod \"certified-operators-llmzg\" (UID: \"e0220445-630a-4d62-91c9-ba260e6ba945\") " pod="openshift-marketplace/certified-operators-llmzg" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.841043 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0220445-630a-4d62-91c9-ba260e6ba945-utilities\") pod \"certified-operators-llmzg\" (UID: \"e0220445-630a-4d62-91c9-ba260e6ba945\") " pod="openshift-marketplace/certified-operators-llmzg" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.841442 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0220445-630a-4d62-91c9-ba260e6ba945-catalog-content\") pod \"certified-operators-llmzg\" (UID: \"e0220445-630a-4d62-91c9-ba260e6ba945\") " pod="openshift-marketplace/certified-operators-llmzg" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.864187 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jsjc\" (UniqueName: \"kubernetes.io/projected/e0220445-630a-4d62-91c9-ba260e6ba945-kube-api-access-2jsjc\") pod \"certified-operators-llmzg\" (UID: \"e0220445-630a-4d62-91c9-ba260e6ba945\") " pod="openshift-marketplace/certified-operators-llmzg" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.894961 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-llmzg" Oct 02 09:50:56 crc kubenswrapper[4934]: I1002 09:50:56.930044 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.001382 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fj8zn"] Oct 02 09:50:57 crc kubenswrapper[4934]: W1002 09:50:57.015711 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab1258e2_b9fc_44da_8e7d_c9e70cd7897d.slice/crio-80feceba4fa69796d9846e77171252701feadf88972ac15345990b7d18b90b32 WatchSource:0}: Error finding container 80feceba4fa69796d9846e77171252701feadf88972ac15345990b7d18b90b32: Status 404 returned error can't find the container with id 80feceba4fa69796d9846e77171252701feadf88972ac15345990b7d18b90b32 Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.019913 4934 patch_prober.go:28] interesting pod/router-default-5444994796-wjz5g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:50:57 crc kubenswrapper[4934]: [-]has-synced failed: reason withheld Oct 02 09:50:57 crc kubenswrapper[4934]: [+]process-running ok Oct 02 09:50:57 crc kubenswrapper[4934]: healthz check failed Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.019959 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wjz5g" podUID="8391ee1a-9c57-4745-825f-381f47254807" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.099141 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-llmzg"] Oct 02 09:50:57 crc kubenswrapper[4934]: W1002 09:50:57.114265 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0220445_630a_4d62_91c9_ba260e6ba945.slice/crio-24c0f821815b6b1821c6c9866e3b95f9178b900d9b4165882b0550fbb52d0d4a WatchSource:0}: Error finding container 24c0f821815b6b1821c6c9866e3b95f9178b900d9b4165882b0550fbb52d0d4a: Status 404 returned error can't find the container with id 24c0f821815b6b1821c6c9866e3b95f9178b900d9b4165882b0550fbb52d0d4a Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.181714 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.187636 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-fc4l5" Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.564846 4934 generic.go:334] "Generic (PLEG): container finished" podID="c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a" containerID="0de8328c21c16ce1160c7e885fa77ebef1fa96de06d43610aad59115c78bf81e" exitCode=0 Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.564937 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-jv64v" event={"ID":"c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a","Type":"ContainerDied","Data":"0de8328c21c16ce1160c7e885fa77ebef1fa96de06d43610aad59115c78bf81e"} Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.567058 4934 generic.go:334] "Generic (PLEG): container finished" podID="ab1258e2-b9fc-44da-8e7d-c9e70cd7897d" containerID="fcd791c407cfe2fc00e7b675bb48c2eee045ee451b731b40ef99a229ced72ae9" exitCode=0 Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.567126 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fj8zn" event={"ID":"ab1258e2-b9fc-44da-8e7d-c9e70cd7897d","Type":"ContainerDied","Data":"fcd791c407cfe2fc00e7b675bb48c2eee045ee451b731b40ef99a229ced72ae9"} Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.567152 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fj8zn" event={"ID":"ab1258e2-b9fc-44da-8e7d-c9e70cd7897d","Type":"ContainerStarted","Data":"80feceba4fa69796d9846e77171252701feadf88972ac15345990b7d18b90b32"} Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.569415 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.573530 4934 generic.go:334] "Generic (PLEG): container finished" podID="ea34a4e8-c842-4a23-933a-8ecd3f882d19" containerID="d1ed750ce3a57c63d62725b7606b9bc3a083a20ae715646eadbfa54479a8b66b" exitCode=0 Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.573631 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6k6dn" event={"ID":"ea34a4e8-c842-4a23-933a-8ecd3f882d19","Type":"ContainerDied","Data":"d1ed750ce3a57c63d62725b7606b9bc3a083a20ae715646eadbfa54479a8b66b"} Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.573660 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6k6dn" event={"ID":"ea34a4e8-c842-4a23-933a-8ecd3f882d19","Type":"ContainerStarted","Data":"755a3ce9f22b3c84ce99df5512871cb51b4c4c06906ce8a9b9fc7f18c6463e38"} Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.576193 4934 generic.go:334] "Generic (PLEG): container finished" podID="e0220445-630a-4d62-91c9-ba260e6ba945" containerID="e8937a6c4cedc22c9713537c06d00a9b28cba540f381db5bc4ef5ce0e74a5313" exitCode=0 Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.576239 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-llmzg" event={"ID":"e0220445-630a-4d62-91c9-ba260e6ba945","Type":"ContainerDied","Data":"e8937a6c4cedc22c9713537c06d00a9b28cba540f381db5bc4ef5ce0e74a5313"} Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.576258 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-llmzg" event={"ID":"e0220445-630a-4d62-91c9-ba260e6ba945","Type":"ContainerStarted","Data":"24c0f821815b6b1821c6c9866e3b95f9178b900d9b4165882b0550fbb52d0d4a"} Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.578706 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" event={"ID":"39e0228e-f4f6-4167-9bbc-d319374345a1","Type":"ContainerStarted","Data":"fa4e833287fc888595e5a530be3afd339fb3ef224dafb04c4032ae16485aef73"} Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.578726 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" event={"ID":"39e0228e-f4f6-4167-9bbc-d319374345a1","Type":"ContainerStarted","Data":"9c6741c48da97be054129169423fe482a7880193c8c331dd0144ab1ae4dba703"} Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.578879 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.580624 4934 generic.go:334] "Generic (PLEG): container finished" podID="b8150488-b2c9-45e0-8209-9d6a73b64834" containerID="8c7dd1fdadb86aa4fdd32321fb249a4eb56f48f5b7388f0a5c00ea5595ec3dd6" exitCode=0 Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.580776 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fdrpb" event={"ID":"b8150488-b2c9-45e0-8209-9d6a73b64834","Type":"ContainerDied","Data":"8c7dd1fdadb86aa4fdd32321fb249a4eb56f48f5b7388f0a5c00ea5595ec3dd6"} Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.580843 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fdrpb" event={"ID":"b8150488-b2c9-45e0-8209-9d6a73b64834","Type":"ContainerStarted","Data":"9fd017cdcae70a0f399ea6316cbbfec49f9b1b62aa48164dd181dc43b6172807"} Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.658786 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" podStartSLOduration=107.658769421 podStartE2EDuration="1m47.658769421s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:50:57.630505179 +0000 UTC m=+129.383146721" watchObservedRunningTime="2025-10-02 09:50:57.658769421 +0000 UTC m=+129.411410943" Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.838568 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.946836 4934 patch_prober.go:28] interesting pod/downloads-7954f5f757-mcnd2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.946930 4934 patch_prober.go:28] interesting pod/downloads-7954f5f757-mcnd2 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.947533 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-mcnd2" podUID="5dbb4e0e-9898-44ab-81f9-d149b80cfef0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.947448 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mcnd2" podUID="5dbb4e0e-9898-44ab-81f9-d149b80cfef0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.957271 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb154e95-c062-4838-ba59-288e60c9138d-kube-api-access\") pod \"eb154e95-c062-4838-ba59-288e60c9138d\" (UID: \"eb154e95-c062-4838-ba59-288e60c9138d\") " Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.957429 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/eb154e95-c062-4838-ba59-288e60c9138d-kubelet-dir\") pod \"eb154e95-c062-4838-ba59-288e60c9138d\" (UID: \"eb154e95-c062-4838-ba59-288e60c9138d\") " Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.957860 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eb154e95-c062-4838-ba59-288e60c9138d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "eb154e95-c062-4838-ba59-288e60c9138d" (UID: "eb154e95-c062-4838-ba59-288e60c9138d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:50:57 crc kubenswrapper[4934]: I1002 09:50:57.968452 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb154e95-c062-4838-ba59-288e60c9138d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "eb154e95-c062-4838-ba59-288e60c9138d" (UID: "eb154e95-c062-4838-ba59-288e60c9138d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.017648 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-wjz5g" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.025094 4934 patch_prober.go:28] interesting pod/router-default-5444994796-wjz5g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:50:58 crc kubenswrapper[4934]: [-]has-synced failed: reason withheld Oct 02 09:50:58 crc kubenswrapper[4934]: [+]process-running ok Oct 02 09:50:58 crc kubenswrapper[4934]: healthz check failed Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.025180 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wjz5g" podUID="8391ee1a-9c57-4745-825f-381f47254807" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.059136 4934 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/eb154e95-c062-4838-ba59-288e60c9138d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.059193 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb154e95-c062-4838-ba59-288e60c9138d-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.152501 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-78cmv"] Oct 02 09:50:58 crc kubenswrapper[4934]: E1002 09:50:58.152931 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb154e95-c062-4838-ba59-288e60c9138d" containerName="pruner" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.152944 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb154e95-c062-4838-ba59-288e60c9138d" containerName="pruner" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.153046 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb154e95-c062-4838-ba59-288e60c9138d" containerName="pruner" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.153878 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-78cmv" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.155484 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.162192 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-78cmv"] Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.199360 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-f44r4" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.199417 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-f44r4" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.201983 4934 patch_prober.go:28] interesting pod/console-f9d7485db-f44r4 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.28:8443/health\": dial tcp 10.217.0.28:8443: connect: connection refused" start-of-body= Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.202026 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-f44r4" podUID="ad155aa0-50a3-4874-93bd-2166a8d093cb" containerName="console" probeResult="failure" output="Get \"https://10.217.0.28:8443/health\": dial tcp 10.217.0.28:8443: connect: connection refused" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.260971 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qgh9\" (UniqueName: \"kubernetes.io/projected/fac56d4f-c941-486b-93cb-6931afbd89e0-kube-api-access-2qgh9\") pod \"redhat-marketplace-78cmv\" (UID: \"fac56d4f-c941-486b-93cb-6931afbd89e0\") " pod="openshift-marketplace/redhat-marketplace-78cmv" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.261018 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fac56d4f-c941-486b-93cb-6931afbd89e0-catalog-content\") pod \"redhat-marketplace-78cmv\" (UID: \"fac56d4f-c941-486b-93cb-6931afbd89e0\") " pod="openshift-marketplace/redhat-marketplace-78cmv" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.261341 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fac56d4f-c941-486b-93cb-6931afbd89e0-utilities\") pod \"redhat-marketplace-78cmv\" (UID: \"fac56d4f-c941-486b-93cb-6931afbd89e0\") " pod="openshift-marketplace/redhat-marketplace-78cmv" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.268459 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.337708 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-66hrr" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.362696 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fac56d4f-c941-486b-93cb-6931afbd89e0-utilities\") pod \"redhat-marketplace-78cmv\" (UID: \"fac56d4f-c941-486b-93cb-6931afbd89e0\") " pod="openshift-marketplace/redhat-marketplace-78cmv" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.362808 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qgh9\" (UniqueName: \"kubernetes.io/projected/fac56d4f-c941-486b-93cb-6931afbd89e0-kube-api-access-2qgh9\") pod \"redhat-marketplace-78cmv\" (UID: \"fac56d4f-c941-486b-93cb-6931afbd89e0\") " pod="openshift-marketplace/redhat-marketplace-78cmv" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.362843 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fac56d4f-c941-486b-93cb-6931afbd89e0-catalog-content\") pod \"redhat-marketplace-78cmv\" (UID: \"fac56d4f-c941-486b-93cb-6931afbd89e0\") " pod="openshift-marketplace/redhat-marketplace-78cmv" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.364444 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fac56d4f-c941-486b-93cb-6931afbd89e0-catalog-content\") pod \"redhat-marketplace-78cmv\" (UID: \"fac56d4f-c941-486b-93cb-6931afbd89e0\") " pod="openshift-marketplace/redhat-marketplace-78cmv" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.365018 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fac56d4f-c941-486b-93cb-6931afbd89e0-utilities\") pod \"redhat-marketplace-78cmv\" (UID: \"fac56d4f-c941-486b-93cb-6931afbd89e0\") " pod="openshift-marketplace/redhat-marketplace-78cmv" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.396272 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qgh9\" (UniqueName: \"kubernetes.io/projected/fac56d4f-c941-486b-93cb-6931afbd89e0-kube-api-access-2qgh9\") pod \"redhat-marketplace-78cmv\" (UID: \"fac56d4f-c941-486b-93cb-6931afbd89e0\") " pod="openshift-marketplace/redhat-marketplace-78cmv" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.470860 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-78cmv" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.553122 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mz6xf"] Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.573353 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mz6xf" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.574969 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mz6xf"] Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.618365 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.618474 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"eb154e95-c062-4838-ba59-288e60c9138d","Type":"ContainerDied","Data":"25fb10aaba38cb2c0c5d4bd4ec40b70b80e2ec9a91f54c9567afbb14fd1fc743"} Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.620663 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25fb10aaba38cb2c0c5d4bd4ec40b70b80e2ec9a91f54c9567afbb14fd1fc743" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.676952 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7c3f392-8c78-4a39-ac9f-d16cb2832698-catalog-content\") pod \"redhat-marketplace-mz6xf\" (UID: \"a7c3f392-8c78-4a39-ac9f-d16cb2832698\") " pod="openshift-marketplace/redhat-marketplace-mz6xf" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.677047 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkgwt\" (UniqueName: \"kubernetes.io/projected/a7c3f392-8c78-4a39-ac9f-d16cb2832698-kube-api-access-jkgwt\") pod \"redhat-marketplace-mz6xf\" (UID: \"a7c3f392-8c78-4a39-ac9f-d16cb2832698\") " pod="openshift-marketplace/redhat-marketplace-mz6xf" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.677075 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7c3f392-8c78-4a39-ac9f-d16cb2832698-utilities\") pod \"redhat-marketplace-mz6xf\" (UID: \"a7c3f392-8c78-4a39-ac9f-d16cb2832698\") " pod="openshift-marketplace/redhat-marketplace-mz6xf" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.777914 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7c3f392-8c78-4a39-ac9f-d16cb2832698-catalog-content\") pod \"redhat-marketplace-mz6xf\" (UID: \"a7c3f392-8c78-4a39-ac9f-d16cb2832698\") " pod="openshift-marketplace/redhat-marketplace-mz6xf" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.778069 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkgwt\" (UniqueName: \"kubernetes.io/projected/a7c3f392-8c78-4a39-ac9f-d16cb2832698-kube-api-access-jkgwt\") pod \"redhat-marketplace-mz6xf\" (UID: \"a7c3f392-8c78-4a39-ac9f-d16cb2832698\") " pod="openshift-marketplace/redhat-marketplace-mz6xf" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.778104 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7c3f392-8c78-4a39-ac9f-d16cb2832698-utilities\") pod \"redhat-marketplace-mz6xf\" (UID: \"a7c3f392-8c78-4a39-ac9f-d16cb2832698\") " pod="openshift-marketplace/redhat-marketplace-mz6xf" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.779545 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7c3f392-8c78-4a39-ac9f-d16cb2832698-catalog-content\") pod \"redhat-marketplace-mz6xf\" (UID: \"a7c3f392-8c78-4a39-ac9f-d16cb2832698\") " pod="openshift-marketplace/redhat-marketplace-mz6xf" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.779808 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7c3f392-8c78-4a39-ac9f-d16cb2832698-utilities\") pod \"redhat-marketplace-mz6xf\" (UID: \"a7c3f392-8c78-4a39-ac9f-d16cb2832698\") " pod="openshift-marketplace/redhat-marketplace-mz6xf" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.814807 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkgwt\" (UniqueName: \"kubernetes.io/projected/a7c3f392-8c78-4a39-ac9f-d16cb2832698-kube-api-access-jkgwt\") pod \"redhat-marketplace-mz6xf\" (UID: \"a7c3f392-8c78-4a39-ac9f-d16cb2832698\") " pod="openshift-marketplace/redhat-marketplace-mz6xf" Oct 02 09:50:58 crc kubenswrapper[4934]: I1002 09:50:58.924811 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mz6xf" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.017798 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-78cmv"] Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.024455 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-wjz5g" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.033453 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-wjz5g" Oct 02 09:50:59 crc kubenswrapper[4934]: W1002 09:50:59.069991 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfac56d4f_c941_486b_93cb_6931afbd89e0.slice/crio-0558673d11c10c65ad87f8aa4d3c4277743949fc1e3a8915af0d31fbaca77fa3 WatchSource:0}: Error finding container 0558673d11c10c65ad87f8aa4d3c4277743949fc1e3a8915af0d31fbaca77fa3: Status 404 returned error can't find the container with id 0558673d11c10c65ad87f8aa4d3c4277743949fc1e3a8915af0d31fbaca77fa3 Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.161231 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-jv64v" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.180388 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vn2xd"] Oct 02 09:50:59 crc kubenswrapper[4934]: E1002 09:50:59.180625 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a" containerName="collect-profiles" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.180637 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a" containerName="collect-profiles" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.180741 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a" containerName="collect-profiles" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.181462 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vn2xd" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.183361 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.206007 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vn2xd"] Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.247196 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.247932 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.250436 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.254562 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.286704 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsgsr\" (UniqueName: \"kubernetes.io/projected/c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a-kube-api-access-vsgsr\") pod \"c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a\" (UID: \"c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a\") " Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.286811 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a-secret-volume\") pod \"c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a\" (UID: \"c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a\") " Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.286850 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a-config-volume\") pod \"c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a\" (UID: \"c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a\") " Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.286991 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acf903d9-ba78-488a-b55e-6ecd290fe774-utilities\") pod \"redhat-operators-vn2xd\" (UID: \"acf903d9-ba78-488a-b55e-6ecd290fe774\") " pod="openshift-marketplace/redhat-operators-vn2xd" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.287011 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsmhj\" (UniqueName: \"kubernetes.io/projected/acf903d9-ba78-488a-b55e-6ecd290fe774-kube-api-access-dsmhj\") pod \"redhat-operators-vn2xd\" (UID: \"acf903d9-ba78-488a-b55e-6ecd290fe774\") " pod="openshift-marketplace/redhat-operators-vn2xd" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.287096 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acf903d9-ba78-488a-b55e-6ecd290fe774-catalog-content\") pod \"redhat-operators-vn2xd\" (UID: \"acf903d9-ba78-488a-b55e-6ecd290fe774\") " pod="openshift-marketplace/redhat-operators-vn2xd" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.288038 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a-config-volume" (OuterVolumeSpecName: "config-volume") pod "c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a" (UID: "c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.295779 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a-kube-api-access-vsgsr" (OuterVolumeSpecName: "kube-api-access-vsgsr") pod "c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a" (UID: "c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a"). InnerVolumeSpecName "kube-api-access-vsgsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.295997 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a" (UID: "c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.314280 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.389742 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acf903d9-ba78-488a-b55e-6ecd290fe774-utilities\") pod \"redhat-operators-vn2xd\" (UID: \"acf903d9-ba78-488a-b55e-6ecd290fe774\") " pod="openshift-marketplace/redhat-operators-vn2xd" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.389798 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsmhj\" (UniqueName: \"kubernetes.io/projected/acf903d9-ba78-488a-b55e-6ecd290fe774-kube-api-access-dsmhj\") pod \"redhat-operators-vn2xd\" (UID: \"acf903d9-ba78-488a-b55e-6ecd290fe774\") " pod="openshift-marketplace/redhat-operators-vn2xd" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.389901 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acf903d9-ba78-488a-b55e-6ecd290fe774-catalog-content\") pod \"redhat-operators-vn2xd\" (UID: \"acf903d9-ba78-488a-b55e-6ecd290fe774\") " pod="openshift-marketplace/redhat-operators-vn2xd" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.402771 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a0c99a28-1f5b-4a44-a03c-e9c3b3254154-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"a0c99a28-1f5b-4a44-a03c-e9c3b3254154\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.402884 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a0c99a28-1f5b-4a44-a03c-e9c3b3254154-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"a0c99a28-1f5b-4a44-a03c-e9c3b3254154\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.402957 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsgsr\" (UniqueName: \"kubernetes.io/projected/c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a-kube-api-access-vsgsr\") on node \"crc\" DevicePath \"\"" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.402969 4934 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.402978 4934 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.390535 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acf903d9-ba78-488a-b55e-6ecd290fe774-catalog-content\") pod \"redhat-operators-vn2xd\" (UID: \"acf903d9-ba78-488a-b55e-6ecd290fe774\") " pod="openshift-marketplace/redhat-operators-vn2xd" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.412349 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acf903d9-ba78-488a-b55e-6ecd290fe774-utilities\") pod \"redhat-operators-vn2xd\" (UID: \"acf903d9-ba78-488a-b55e-6ecd290fe774\") " pod="openshift-marketplace/redhat-operators-vn2xd" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.415647 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsmhj\" (UniqueName: \"kubernetes.io/projected/acf903d9-ba78-488a-b55e-6ecd290fe774-kube-api-access-dsmhj\") pod \"redhat-operators-vn2xd\" (UID: \"acf903d9-ba78-488a-b55e-6ecd290fe774\") " pod="openshift-marketplace/redhat-operators-vn2xd" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.504263 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a0c99a28-1f5b-4a44-a03c-e9c3b3254154-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"a0c99a28-1f5b-4a44-a03c-e9c3b3254154\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.504636 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a0c99a28-1f5b-4a44-a03c-e9c3b3254154-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"a0c99a28-1f5b-4a44-a03c-e9c3b3254154\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.504702 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a0c99a28-1f5b-4a44-a03c-e9c3b3254154-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"a0c99a28-1f5b-4a44-a03c-e9c3b3254154\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.512952 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vn2xd" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.547270 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a0c99a28-1f5b-4a44-a03c-e9c3b3254154-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"a0c99a28-1f5b-4a44-a03c-e9c3b3254154\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.558884 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lrg5q"] Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.559908 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lrg5q" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.573118 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.574121 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lrg5q"] Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.658227 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-78cmv" event={"ID":"fac56d4f-c941-486b-93cb-6931afbd89e0","Type":"ContainerStarted","Data":"0558673d11c10c65ad87f8aa4d3c4277743949fc1e3a8915af0d31fbaca77fa3"} Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.661836 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-jv64v" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.661832 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-jv64v" event={"ID":"c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a","Type":"ContainerDied","Data":"b57e8d69babf273a49c2f378fa8455b2d149d4697bca4d2369bac71a0c97d6e5"} Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.661891 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b57e8d69babf273a49c2f378fa8455b2d149d4697bca4d2369bac71a0c97d6e5" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.719037 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be630036-c659-47b0-a7bb-bac356b7df7a-catalog-content\") pod \"redhat-operators-lrg5q\" (UID: \"be630036-c659-47b0-a7bb-bac356b7df7a\") " pod="openshift-marketplace/redhat-operators-lrg5q" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.719177 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xv2bp\" (UniqueName: \"kubernetes.io/projected/be630036-c659-47b0-a7bb-bac356b7df7a-kube-api-access-xv2bp\") pod \"redhat-operators-lrg5q\" (UID: \"be630036-c659-47b0-a7bb-bac356b7df7a\") " pod="openshift-marketplace/redhat-operators-lrg5q" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.719232 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be630036-c659-47b0-a7bb-bac356b7df7a-utilities\") pod \"redhat-operators-lrg5q\" (UID: \"be630036-c659-47b0-a7bb-bac356b7df7a\") " pod="openshift-marketplace/redhat-operators-lrg5q" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.729338 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mz6xf"] Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.822763 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be630036-c659-47b0-a7bb-bac356b7df7a-utilities\") pod \"redhat-operators-lrg5q\" (UID: \"be630036-c659-47b0-a7bb-bac356b7df7a\") " pod="openshift-marketplace/redhat-operators-lrg5q" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.822902 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be630036-c659-47b0-a7bb-bac356b7df7a-catalog-content\") pod \"redhat-operators-lrg5q\" (UID: \"be630036-c659-47b0-a7bb-bac356b7df7a\") " pod="openshift-marketplace/redhat-operators-lrg5q" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.823015 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xv2bp\" (UniqueName: \"kubernetes.io/projected/be630036-c659-47b0-a7bb-bac356b7df7a-kube-api-access-xv2bp\") pod \"redhat-operators-lrg5q\" (UID: \"be630036-c659-47b0-a7bb-bac356b7df7a\") " pod="openshift-marketplace/redhat-operators-lrg5q" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.825182 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be630036-c659-47b0-a7bb-bac356b7df7a-utilities\") pod \"redhat-operators-lrg5q\" (UID: \"be630036-c659-47b0-a7bb-bac356b7df7a\") " pod="openshift-marketplace/redhat-operators-lrg5q" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.825188 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be630036-c659-47b0-a7bb-bac356b7df7a-catalog-content\") pod \"redhat-operators-lrg5q\" (UID: \"be630036-c659-47b0-a7bb-bac356b7df7a\") " pod="openshift-marketplace/redhat-operators-lrg5q" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.849916 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xv2bp\" (UniqueName: \"kubernetes.io/projected/be630036-c659-47b0-a7bb-bac356b7df7a-kube-api-access-xv2bp\") pod \"redhat-operators-lrg5q\" (UID: \"be630036-c659-47b0-a7bb-bac356b7df7a\") " pod="openshift-marketplace/redhat-operators-lrg5q" Oct 02 09:50:59 crc kubenswrapper[4934]: I1002 09:50:59.885367 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lrg5q" Oct 02 09:51:00 crc kubenswrapper[4934]: I1002 09:51:00.054480 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 09:51:00 crc kubenswrapper[4934]: W1002 09:51:00.065141 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poda0c99a28_1f5b_4a44_a03c_e9c3b3254154.slice/crio-cab6fb9fdb1fb05d9ee61f59fe0d6814f82a969972828a7abc0476b0637ea811 WatchSource:0}: Error finding container cab6fb9fdb1fb05d9ee61f59fe0d6814f82a969972828a7abc0476b0637ea811: Status 404 returned error can't find the container with id cab6fb9fdb1fb05d9ee61f59fe0d6814f82a969972828a7abc0476b0637ea811 Oct 02 09:51:00 crc kubenswrapper[4934]: I1002 09:51:00.205513 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vn2xd"] Oct 02 09:51:00 crc kubenswrapper[4934]: I1002 09:51:00.239595 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lrg5q"] Oct 02 09:51:00 crc kubenswrapper[4934]: W1002 09:51:00.298873 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe630036_c659_47b0_a7bb_bac356b7df7a.slice/crio-3083ae48699cfb77960ebfffa8d17f427a14e293ae41fa0b0b97c4f9338cfadf WatchSource:0}: Error finding container 3083ae48699cfb77960ebfffa8d17f427a14e293ae41fa0b0b97c4f9338cfadf: Status 404 returned error can't find the container with id 3083ae48699cfb77960ebfffa8d17f427a14e293ae41fa0b0b97c4f9338cfadf Oct 02 09:51:00 crc kubenswrapper[4934]: I1002 09:51:00.692191 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vn2xd" event={"ID":"acf903d9-ba78-488a-b55e-6ecd290fe774","Type":"ContainerStarted","Data":"1a6331787b0998c2756ee4ca02b34e4d188a6420ff77a517022e7649f87920ae"} Oct 02 09:51:00 crc kubenswrapper[4934]: I1002 09:51:00.700852 4934 generic.go:334] "Generic (PLEG): container finished" podID="a7c3f392-8c78-4a39-ac9f-d16cb2832698" containerID="24bf0e3bafa23c98917c42b27fc71fffee12d56d5afd06458a775891755e38ac" exitCode=0 Oct 02 09:51:00 crc kubenswrapper[4934]: I1002 09:51:00.700962 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mz6xf" event={"ID":"a7c3f392-8c78-4a39-ac9f-d16cb2832698","Type":"ContainerDied","Data":"24bf0e3bafa23c98917c42b27fc71fffee12d56d5afd06458a775891755e38ac"} Oct 02 09:51:00 crc kubenswrapper[4934]: I1002 09:51:00.700997 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mz6xf" event={"ID":"a7c3f392-8c78-4a39-ac9f-d16cb2832698","Type":"ContainerStarted","Data":"ee165fb3416e086e6dd7dfca2c37aad15fc78d246964d0bfbf1ef8d80c961445"} Oct 02 09:51:00 crc kubenswrapper[4934]: I1002 09:51:00.703652 4934 generic.go:334] "Generic (PLEG): container finished" podID="fac56d4f-c941-486b-93cb-6931afbd89e0" containerID="7b7e05d4289a96d7ab8c886aae97669defac87fa69c970a55928de92875c792d" exitCode=0 Oct 02 09:51:00 crc kubenswrapper[4934]: I1002 09:51:00.703694 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-78cmv" event={"ID":"fac56d4f-c941-486b-93cb-6931afbd89e0","Type":"ContainerDied","Data":"7b7e05d4289a96d7ab8c886aae97669defac87fa69c970a55928de92875c792d"} Oct 02 09:51:00 crc kubenswrapper[4934]: I1002 09:51:00.706236 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"a0c99a28-1f5b-4a44-a03c-e9c3b3254154","Type":"ContainerStarted","Data":"cab6fb9fdb1fb05d9ee61f59fe0d6814f82a969972828a7abc0476b0637ea811"} Oct 02 09:51:00 crc kubenswrapper[4934]: I1002 09:51:00.712626 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrg5q" event={"ID":"be630036-c659-47b0-a7bb-bac356b7df7a","Type":"ContainerStarted","Data":"3083ae48699cfb77960ebfffa8d17f427a14e293ae41fa0b0b97c4f9338cfadf"} Oct 02 09:51:01 crc kubenswrapper[4934]: I1002 09:51:01.726034 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"a0c99a28-1f5b-4a44-a03c-e9c3b3254154","Type":"ContainerStarted","Data":"d8bc7019c23d1ca0744393ae3dcba592c7a42ec9524c40596fc442d8d6469531"} Oct 02 09:51:01 crc kubenswrapper[4934]: I1002 09:51:01.735422 4934 generic.go:334] "Generic (PLEG): container finished" podID="be630036-c659-47b0-a7bb-bac356b7df7a" containerID="3954675970c95e9b2f83cbd67321838b462d4590c2c82e624a4e51f1586973e4" exitCode=0 Oct 02 09:51:01 crc kubenswrapper[4934]: I1002 09:51:01.735588 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrg5q" event={"ID":"be630036-c659-47b0-a7bb-bac356b7df7a","Type":"ContainerDied","Data":"3954675970c95e9b2f83cbd67321838b462d4590c2c82e624a4e51f1586973e4"} Oct 02 09:51:01 crc kubenswrapper[4934]: I1002 09:51:01.739037 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.739020903 podStartE2EDuration="2.739020903s" podCreationTimestamp="2025-10-02 09:50:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:51:01.737980806 +0000 UTC m=+133.490622348" watchObservedRunningTime="2025-10-02 09:51:01.739020903 +0000 UTC m=+133.491662425" Oct 02 09:51:01 crc kubenswrapper[4934]: I1002 09:51:01.751993 4934 generic.go:334] "Generic (PLEG): container finished" podID="acf903d9-ba78-488a-b55e-6ecd290fe774" containerID="e8bb383ed717e61065ad617fbe6a0ee34e1f8a93c8eb038696f510bbdee61dd9" exitCode=0 Oct 02 09:51:01 crc kubenswrapper[4934]: I1002 09:51:01.752230 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vn2xd" event={"ID":"acf903d9-ba78-488a-b55e-6ecd290fe774","Type":"ContainerDied","Data":"e8bb383ed717e61065ad617fbe6a0ee34e1f8a93c8eb038696f510bbdee61dd9"} Oct 02 09:51:02 crc kubenswrapper[4934]: I1002 09:51:02.767059 4934 generic.go:334] "Generic (PLEG): container finished" podID="a0c99a28-1f5b-4a44-a03c-e9c3b3254154" containerID="d8bc7019c23d1ca0744393ae3dcba592c7a42ec9524c40596fc442d8d6469531" exitCode=0 Oct 02 09:51:02 crc kubenswrapper[4934]: I1002 09:51:02.767113 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"a0c99a28-1f5b-4a44-a03c-e9c3b3254154","Type":"ContainerDied","Data":"d8bc7019c23d1ca0744393ae3dcba592c7a42ec9524c40596fc442d8d6469531"} Oct 02 09:51:03 crc kubenswrapper[4934]: I1002 09:51:03.492781 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-v58f6" Oct 02 09:51:07 crc kubenswrapper[4934]: I1002 09:51:07.953707 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-mcnd2" Oct 02 09:51:08 crc kubenswrapper[4934]: I1002 09:51:08.203905 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-f44r4" Oct 02 09:51:08 crc kubenswrapper[4934]: I1002 09:51:08.208467 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-f44r4" Oct 02 09:51:10 crc kubenswrapper[4934]: I1002 09:51:10.585337 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:51:10 crc kubenswrapper[4934]: I1002 09:51:10.675882 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 09:51:10 crc kubenswrapper[4934]: I1002 09:51:10.783536 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a0c99a28-1f5b-4a44-a03c-e9c3b3254154-kube-api-access\") pod \"a0c99a28-1f5b-4a44-a03c-e9c3b3254154\" (UID: \"a0c99a28-1f5b-4a44-a03c-e9c3b3254154\") " Oct 02 09:51:10 crc kubenswrapper[4934]: I1002 09:51:10.783901 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a0c99a28-1f5b-4a44-a03c-e9c3b3254154-kubelet-dir\") pod \"a0c99a28-1f5b-4a44-a03c-e9c3b3254154\" (UID: \"a0c99a28-1f5b-4a44-a03c-e9c3b3254154\") " Oct 02 09:51:10 crc kubenswrapper[4934]: I1002 09:51:10.784158 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a0c99a28-1f5b-4a44-a03c-e9c3b3254154-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a0c99a28-1f5b-4a44-a03c-e9c3b3254154" (UID: "a0c99a28-1f5b-4a44-a03c-e9c3b3254154"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:51:10 crc kubenswrapper[4934]: I1002 09:51:10.789435 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0c99a28-1f5b-4a44-a03c-e9c3b3254154-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a0c99a28-1f5b-4a44-a03c-e9c3b3254154" (UID: "a0c99a28-1f5b-4a44-a03c-e9c3b3254154"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:51:10 crc kubenswrapper[4934]: I1002 09:51:10.823822 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"a0c99a28-1f5b-4a44-a03c-e9c3b3254154","Type":"ContainerDied","Data":"cab6fb9fdb1fb05d9ee61f59fe0d6814f82a969972828a7abc0476b0637ea811"} Oct 02 09:51:10 crc kubenswrapper[4934]: I1002 09:51:10.823864 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cab6fb9fdb1fb05d9ee61f59fe0d6814f82a969972828a7abc0476b0637ea811" Oct 02 09:51:10 crc kubenswrapper[4934]: I1002 09:51:10.823940 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:51:10 crc kubenswrapper[4934]: I1002 09:51:10.884926 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a0c99a28-1f5b-4a44-a03c-e9c3b3254154-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 09:51:10 crc kubenswrapper[4934]: I1002 09:51:10.884966 4934 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a0c99a28-1f5b-4a44-a03c-e9c3b3254154-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 02 09:51:16 crc kubenswrapper[4934]: I1002 09:51:16.070285 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:51:16 crc kubenswrapper[4934]: I1002 09:51:16.073065 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 02 09:51:16 crc kubenswrapper[4934]: I1002 09:51:16.086113 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:51:16 crc kubenswrapper[4934]: I1002 09:51:16.171664 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:51:16 crc kubenswrapper[4934]: I1002 09:51:16.171748 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:51:16 crc kubenswrapper[4934]: I1002 09:51:16.171797 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:51:16 crc kubenswrapper[4934]: I1002 09:51:16.173673 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 02 09:51:16 crc kubenswrapper[4934]: I1002 09:51:16.173786 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 02 09:51:16 crc kubenswrapper[4934]: I1002 09:51:16.183524 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 02 09:51:16 crc kubenswrapper[4934]: I1002 09:51:16.189430 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:51:16 crc kubenswrapper[4934]: I1002 09:51:16.195936 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:51:16 crc kubenswrapper[4934]: I1002 09:51:16.195991 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:51:16 crc kubenswrapper[4934]: I1002 09:51:16.227916 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:51:16 crc kubenswrapper[4934]: I1002 09:51:16.235617 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:51:16 crc kubenswrapper[4934]: I1002 09:51:16.244443 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:51:16 crc kubenswrapper[4934]: I1002 09:51:16.405536 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:51:24 crc kubenswrapper[4934]: E1002 09:51:24.718717 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: writing blob: storing blob to file \"/var/tmp/container_images_storage2779622482/2\": happened during read: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 02 09:51:24 crc kubenswrapper[4934]: E1002 09:51:24.719517 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dsmhj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-vn2xd_openshift-marketplace(acf903d9-ba78-488a-b55e-6ecd290fe774): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: writing blob: storing blob to file \"/var/tmp/container_images_storage2779622482/2\": happened during read: context canceled" logger="UnhandledError" Oct 02 09:51:24 crc kubenswrapper[4934]: E1002 09:51:24.720845 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: writing blob: storing blob to file \\\"/var/tmp/container_images_storage2779622482/2\\\": happened during read: context canceled\"" pod="openshift-marketplace/redhat-operators-vn2xd" podUID="acf903d9-ba78-488a-b55e-6ecd290fe774" Oct 02 09:51:28 crc kubenswrapper[4934]: I1002 09:51:28.696693 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5s688" Oct 02 09:51:29 crc kubenswrapper[4934]: E1002 09:51:29.389842 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-vn2xd" podUID="acf903d9-ba78-488a-b55e-6ecd290fe774" Oct 02 09:51:31 crc kubenswrapper[4934]: E1002 09:51:31.854107 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 02 09:51:31 crc kubenswrapper[4934]: E1002 09:51:31.854428 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-svl4c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-fdrpb_openshift-marketplace(b8150488-b2c9-45e0-8209-9d6a73b64834): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 09:51:31 crc kubenswrapper[4934]: E1002 09:51:31.855592 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-fdrpb" podUID="b8150488-b2c9-45e0-8209-9d6a73b64834" Oct 02 09:51:31 crc kubenswrapper[4934]: I1002 09:51:31.982392 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-metrics-certs\") pod \"network-metrics-daemon-n7cgz\" (UID: \"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\") " pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:51:31 crc kubenswrapper[4934]: I1002 09:51:31.984426 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 02 09:51:32 crc kubenswrapper[4934]: I1002 09:51:32.018441 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0cbef3ee-c66f-47fa-94d3-5ce9892c403f-metrics-certs\") pod \"network-metrics-daemon-n7cgz\" (UID: \"0cbef3ee-c66f-47fa-94d3-5ce9892c403f\") " pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:51:32 crc kubenswrapper[4934]: I1002 09:51:32.237228 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 02 09:51:32 crc kubenswrapper[4934]: I1002 09:51:32.246093 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7cgz" Oct 02 09:51:33 crc kubenswrapper[4934]: E1002 09:51:33.018333 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-fdrpb" podUID="b8150488-b2c9-45e0-8209-9d6a73b64834" Oct 02 09:51:33 crc kubenswrapper[4934]: E1002 09:51:33.086700 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 02 09:51:33 crc kubenswrapper[4934]: E1002 09:51:33.086893 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wj2fk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-fj8zn_openshift-marketplace(ab1258e2-b9fc-44da-8e7d-c9e70cd7897d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 09:51:33 crc kubenswrapper[4934]: E1002 09:51:33.088272 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-fj8zn" podUID="ab1258e2-b9fc-44da-8e7d-c9e70cd7897d" Oct 02 09:51:33 crc kubenswrapper[4934]: E1002 09:51:33.864519 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 02 09:51:33 crc kubenswrapper[4934]: E1002 09:51:33.865362 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jkgwt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-mz6xf_openshift-marketplace(a7c3f392-8c78-4a39-ac9f-d16cb2832698): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 09:51:33 crc kubenswrapper[4934]: E1002 09:51:33.866603 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-mz6xf" podUID="a7c3f392-8c78-4a39-ac9f-d16cb2832698" Oct 02 09:51:33 crc kubenswrapper[4934]: E1002 09:51:33.906274 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 02 09:51:33 crc kubenswrapper[4934]: E1002 09:51:33.906432 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vqmfs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-6k6dn_openshift-marketplace(ea34a4e8-c842-4a23-933a-8ecd3f882d19): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 09:51:33 crc kubenswrapper[4934]: E1002 09:51:33.908152 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-6k6dn" podUID="ea34a4e8-c842-4a23-933a-8ecd3f882d19" Oct 02 09:51:33 crc kubenswrapper[4934]: E1002 09:51:33.912291 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 02 09:51:33 crc kubenswrapper[4934]: E1002 09:51:33.912899 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2qgh9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-78cmv_openshift-marketplace(fac56d4f-c941-486b-93cb-6931afbd89e0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 09:51:33 crc kubenswrapper[4934]: E1002 09:51:33.914864 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-78cmv" podUID="fac56d4f-c941-486b-93cb-6931afbd89e0" Oct 02 09:51:33 crc kubenswrapper[4934]: E1002 09:51:33.950000 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 02 09:51:33 crc kubenswrapper[4934]: E1002 09:51:33.950151 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2jsjc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-llmzg_openshift-marketplace(e0220445-630a-4d62-91c9-ba260e6ba945): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 09:51:33 crc kubenswrapper[4934]: E1002 09:51:33.951245 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-llmzg" podUID="e0220445-630a-4d62-91c9-ba260e6ba945" Oct 02 09:51:36 crc kubenswrapper[4934]: E1002 09:51:36.442475 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-78cmv" podUID="fac56d4f-c941-486b-93cb-6931afbd89e0" Oct 02 09:51:36 crc kubenswrapper[4934]: E1002 09:51:36.442617 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-fj8zn" podUID="ab1258e2-b9fc-44da-8e7d-c9e70cd7897d" Oct 02 09:51:36 crc kubenswrapper[4934]: E1002 09:51:36.442694 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-llmzg" podUID="e0220445-630a-4d62-91c9-ba260e6ba945" Oct 02 09:51:36 crc kubenswrapper[4934]: E1002 09:51:36.442693 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-6k6dn" podUID="ea34a4e8-c842-4a23-933a-8ecd3f882d19" Oct 02 09:51:36 crc kubenswrapper[4934]: I1002 09:51:36.960930 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrg5q" event={"ID":"be630036-c659-47b0-a7bb-bac356b7df7a","Type":"ContainerStarted","Data":"bbbd47eb40cc85d7be63d1285bbe262a125707e1555597419fe066309816b520"} Oct 02 09:51:37 crc kubenswrapper[4934]: W1002 09:51:37.028317 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-c61d00d15038d2e06743f74f41ae4155e69397b883c85b708d65e78c3f5ae4f0 WatchSource:0}: Error finding container c61d00d15038d2e06743f74f41ae4155e69397b883c85b708d65e78c3f5ae4f0: Status 404 returned error can't find the container with id c61d00d15038d2e06743f74f41ae4155e69397b883c85b708d65e78c3f5ae4f0 Oct 02 09:51:37 crc kubenswrapper[4934]: W1002 09:51:37.161188 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-fa9e1dead4a631b6a2b7137696e8064e6487725b89edecd45e2da7502c872fd5 WatchSource:0}: Error finding container fa9e1dead4a631b6a2b7137696e8064e6487725b89edecd45e2da7502c872fd5: Status 404 returned error can't find the container with id fa9e1dead4a631b6a2b7137696e8064e6487725b89edecd45e2da7502c872fd5 Oct 02 09:51:37 crc kubenswrapper[4934]: I1002 09:51:37.167923 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-n7cgz"] Oct 02 09:51:37 crc kubenswrapper[4934]: W1002 09:51:37.182378 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0cbef3ee_c66f_47fa_94d3_5ce9892c403f.slice/crio-9b195f8a6f734892930421f9ad2385fcbcd8a1d62550ac950acc413275af1480 WatchSource:0}: Error finding container 9b195f8a6f734892930421f9ad2385fcbcd8a1d62550ac950acc413275af1480: Status 404 returned error can't find the container with id 9b195f8a6f734892930421f9ad2385fcbcd8a1d62550ac950acc413275af1480 Oct 02 09:51:37 crc kubenswrapper[4934]: I1002 09:51:37.967705 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"4cb7e47da71e514320a5dad101e58743f49b6491871e9192c6aa4baa63217d2e"} Oct 02 09:51:37 crc kubenswrapper[4934]: I1002 09:51:37.968077 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"c61d00d15038d2e06743f74f41ae4155e69397b883c85b708d65e78c3f5ae4f0"} Oct 02 09:51:37 crc kubenswrapper[4934]: I1002 09:51:37.968295 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:51:37 crc kubenswrapper[4934]: I1002 09:51:37.970534 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-n7cgz" event={"ID":"0cbef3ee-c66f-47fa-94d3-5ce9892c403f","Type":"ContainerStarted","Data":"e3548d6f3a33e3d607b839eb0410fc88c6077d6c22fa6896520237e5e6bff060"} Oct 02 09:51:37 crc kubenswrapper[4934]: I1002 09:51:37.970613 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-n7cgz" event={"ID":"0cbef3ee-c66f-47fa-94d3-5ce9892c403f","Type":"ContainerStarted","Data":"238aff5cc5d698f8bfece8bef6c1ccb30396e778e09c859d2186b0f4cb18581d"} Oct 02 09:51:37 crc kubenswrapper[4934]: I1002 09:51:37.970634 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-n7cgz" event={"ID":"0cbef3ee-c66f-47fa-94d3-5ce9892c403f","Type":"ContainerStarted","Data":"9b195f8a6f734892930421f9ad2385fcbcd8a1d62550ac950acc413275af1480"} Oct 02 09:51:37 crc kubenswrapper[4934]: I1002 09:51:37.973370 4934 generic.go:334] "Generic (PLEG): container finished" podID="be630036-c659-47b0-a7bb-bac356b7df7a" containerID="bbbd47eb40cc85d7be63d1285bbe262a125707e1555597419fe066309816b520" exitCode=0 Oct 02 09:51:37 crc kubenswrapper[4934]: I1002 09:51:37.973451 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrg5q" event={"ID":"be630036-c659-47b0-a7bb-bac356b7df7a","Type":"ContainerDied","Data":"bbbd47eb40cc85d7be63d1285bbe262a125707e1555597419fe066309816b520"} Oct 02 09:51:37 crc kubenswrapper[4934]: I1002 09:51:37.976149 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"a1704de84f4a50e25f82038d493da30b39ec8c3743b859f55f8873f9130bf90b"} Oct 02 09:51:37 crc kubenswrapper[4934]: I1002 09:51:37.976182 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"97ae3d2ebba71d313abcb403dac679288a734b5929d37af0b3c0b49315eed39b"} Oct 02 09:51:37 crc kubenswrapper[4934]: I1002 09:51:37.978378 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"2f93bbcacf31d4e91df340f1bd87f9ab6c2043fcf4060b9db79c28b6d29145da"} Oct 02 09:51:37 crc kubenswrapper[4934]: I1002 09:51:37.978422 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"fa9e1dead4a631b6a2b7137696e8064e6487725b89edecd45e2da7502c872fd5"} Oct 02 09:51:38 crc kubenswrapper[4934]: I1002 09:51:38.071367 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-n7cgz" podStartSLOduration=148.071344291 podStartE2EDuration="2m28.071344291s" podCreationTimestamp="2025-10-02 09:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:51:38.054770045 +0000 UTC m=+169.807411597" watchObservedRunningTime="2025-10-02 09:51:38.071344291 +0000 UTC m=+169.823985813" Oct 02 09:51:38 crc kubenswrapper[4934]: I1002 09:51:38.440313 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:51:38 crc kubenswrapper[4934]: I1002 09:51:38.440397 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:51:38 crc kubenswrapper[4934]: I1002 09:51:38.987555 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrg5q" event={"ID":"be630036-c659-47b0-a7bb-bac356b7df7a","Type":"ContainerStarted","Data":"6ea364178614fd8fd8bf29bca4e2adea528f0f02d7ff35f2f0057df68378b81f"} Oct 02 09:51:39 crc kubenswrapper[4934]: I1002 09:51:39.014548 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lrg5q" podStartSLOduration=3.313506619 podStartE2EDuration="40.014500758s" podCreationTimestamp="2025-10-02 09:50:59 +0000 UTC" firstStartedPulling="2025-10-02 09:51:01.7517517 +0000 UTC m=+133.504393222" lastFinishedPulling="2025-10-02 09:51:38.452745839 +0000 UTC m=+170.205387361" observedRunningTime="2025-10-02 09:51:39.008882734 +0000 UTC m=+170.761524256" watchObservedRunningTime="2025-10-02 09:51:39.014500758 +0000 UTC m=+170.767142320" Oct 02 09:51:39 crc kubenswrapper[4934]: I1002 09:51:39.887018 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lrg5q" Oct 02 09:51:39 crc kubenswrapper[4934]: I1002 09:51:39.887090 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lrg5q" Oct 02 09:51:41 crc kubenswrapper[4934]: I1002 09:51:41.068794 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lrg5q" podUID="be630036-c659-47b0-a7bb-bac356b7df7a" containerName="registry-server" probeResult="failure" output=< Oct 02 09:51:41 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Oct 02 09:51:41 crc kubenswrapper[4934]: > Oct 02 09:51:42 crc kubenswrapper[4934]: I1002 09:51:42.003514 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vn2xd" event={"ID":"acf903d9-ba78-488a-b55e-6ecd290fe774","Type":"ContainerStarted","Data":"4b4ab48813919caa119eefa5cc6a5f532f7620db21530457bdd66677157e4f77"} Oct 02 09:51:43 crc kubenswrapper[4934]: I1002 09:51:43.012517 4934 generic.go:334] "Generic (PLEG): container finished" podID="acf903d9-ba78-488a-b55e-6ecd290fe774" containerID="4b4ab48813919caa119eefa5cc6a5f532f7620db21530457bdd66677157e4f77" exitCode=0 Oct 02 09:51:43 crc kubenswrapper[4934]: I1002 09:51:43.012616 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vn2xd" event={"ID":"acf903d9-ba78-488a-b55e-6ecd290fe774","Type":"ContainerDied","Data":"4b4ab48813919caa119eefa5cc6a5f532f7620db21530457bdd66677157e4f77"} Oct 02 09:51:44 crc kubenswrapper[4934]: I1002 09:51:44.018943 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vn2xd" event={"ID":"acf903d9-ba78-488a-b55e-6ecd290fe774","Type":"ContainerStarted","Data":"760adf804fc502d51d5b18c27dad88d647e7fe1ae60a645d8caa715dc86f4341"} Oct 02 09:51:44 crc kubenswrapper[4934]: I1002 09:51:44.938515 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vn2xd" podStartSLOduration=4.27117478 podStartE2EDuration="45.93849145s" podCreationTimestamp="2025-10-02 09:50:59 +0000 UTC" firstStartedPulling="2025-10-02 09:51:01.769012143 +0000 UTC m=+133.521653665" lastFinishedPulling="2025-10-02 09:51:43.436328813 +0000 UTC m=+175.188970335" observedRunningTime="2025-10-02 09:51:44.04003019 +0000 UTC m=+175.792671722" watchObservedRunningTime="2025-10-02 09:51:44.93849145 +0000 UTC m=+176.691132992" Oct 02 09:51:46 crc kubenswrapper[4934]: I1002 09:51:46.031529 4934 generic.go:334] "Generic (PLEG): container finished" podID="a7c3f392-8c78-4a39-ac9f-d16cb2832698" containerID="51ec98110649f0ef94d550060bff4fe3093ec184bfb2dd4d11a76f1252d62763" exitCode=0 Oct 02 09:51:46 crc kubenswrapper[4934]: I1002 09:51:46.031851 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mz6xf" event={"ID":"a7c3f392-8c78-4a39-ac9f-d16cb2832698","Type":"ContainerDied","Data":"51ec98110649f0ef94d550060bff4fe3093ec184bfb2dd4d11a76f1252d62763"} Oct 02 09:51:47 crc kubenswrapper[4934]: I1002 09:51:47.049403 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mz6xf" event={"ID":"a7c3f392-8c78-4a39-ac9f-d16cb2832698","Type":"ContainerStarted","Data":"5713b8fe75a793097de0412c1ed9fdf782e19786308d56da8fa45a3a03ed843e"} Oct 02 09:51:47 crc kubenswrapper[4934]: I1002 09:51:47.070318 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mz6xf" podStartSLOduration=4.13288447 podStartE2EDuration="49.0702933s" podCreationTimestamp="2025-10-02 09:50:58 +0000 UTC" firstStartedPulling="2025-10-02 09:51:01.759660232 +0000 UTC m=+133.512301754" lastFinishedPulling="2025-10-02 09:51:46.697069062 +0000 UTC m=+178.449710584" observedRunningTime="2025-10-02 09:51:47.067606321 +0000 UTC m=+178.820247863" watchObservedRunningTime="2025-10-02 09:51:47.0702933 +0000 UTC m=+178.822934842" Oct 02 09:51:48 crc kubenswrapper[4934]: I1002 09:51:48.925591 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mz6xf" Oct 02 09:51:48 crc kubenswrapper[4934]: I1002 09:51:48.925642 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mz6xf" Oct 02 09:51:49 crc kubenswrapper[4934]: I1002 09:51:49.043379 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mz6xf" Oct 02 09:51:49 crc kubenswrapper[4934]: I1002 09:51:49.513134 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vn2xd" Oct 02 09:51:49 crc kubenswrapper[4934]: I1002 09:51:49.513199 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vn2xd" Oct 02 09:51:49 crc kubenswrapper[4934]: I1002 09:51:49.553993 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vn2xd" Oct 02 09:51:49 crc kubenswrapper[4934]: I1002 09:51:49.932071 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lrg5q" Oct 02 09:51:49 crc kubenswrapper[4934]: I1002 09:51:49.981525 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lrg5q" Oct 02 09:51:50 crc kubenswrapper[4934]: I1002 09:51:50.069848 4934 generic.go:334] "Generic (PLEG): container finished" podID="ab1258e2-b9fc-44da-8e7d-c9e70cd7897d" containerID="647cb41b0117107c7144477c0a14b676be611d916ed38b1dcb7d32433b07a0ee" exitCode=0 Oct 02 09:51:50 crc kubenswrapper[4934]: I1002 09:51:50.069912 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fj8zn" event={"ID":"ab1258e2-b9fc-44da-8e7d-c9e70cd7897d","Type":"ContainerDied","Data":"647cb41b0117107c7144477c0a14b676be611d916ed38b1dcb7d32433b07a0ee"} Oct 02 09:51:50 crc kubenswrapper[4934]: I1002 09:51:50.072995 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6k6dn" event={"ID":"ea34a4e8-c842-4a23-933a-8ecd3f882d19","Type":"ContainerStarted","Data":"3b963f52f5103d064cce85dfc5d3fedb1aa2c7c534722d0be9ba5c4f5278015e"} Oct 02 09:51:50 crc kubenswrapper[4934]: I1002 09:51:50.081727 4934 generic.go:334] "Generic (PLEG): container finished" podID="fac56d4f-c941-486b-93cb-6931afbd89e0" containerID="1ab4526e8d685fd70ffe68ee552c9061d0b65459323e4aea17a8020258448b84" exitCode=0 Oct 02 09:51:50 crc kubenswrapper[4934]: I1002 09:51:50.081813 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-78cmv" event={"ID":"fac56d4f-c941-486b-93cb-6931afbd89e0","Type":"ContainerDied","Data":"1ab4526e8d685fd70ffe68ee552c9061d0b65459323e4aea17a8020258448b84"} Oct 02 09:51:50 crc kubenswrapper[4934]: I1002 09:51:50.085451 4934 generic.go:334] "Generic (PLEG): container finished" podID="b8150488-b2c9-45e0-8209-9d6a73b64834" containerID="f1aea8a08be906a16f7caa992a0e4648b79b5d08dc050ed271c0130f8ebede85" exitCode=0 Oct 02 09:51:50 crc kubenswrapper[4934]: I1002 09:51:50.085544 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fdrpb" event={"ID":"b8150488-b2c9-45e0-8209-9d6a73b64834","Type":"ContainerDied","Data":"f1aea8a08be906a16f7caa992a0e4648b79b5d08dc050ed271c0130f8ebede85"} Oct 02 09:51:50 crc kubenswrapper[4934]: I1002 09:51:50.132063 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vn2xd" Oct 02 09:51:51 crc kubenswrapper[4934]: I1002 09:51:51.094237 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fj8zn" event={"ID":"ab1258e2-b9fc-44da-8e7d-c9e70cd7897d","Type":"ContainerStarted","Data":"f948029f5bf66c725b94a8a7b743870b6258a1dd1ac6482c29f587702a519a61"} Oct 02 09:51:51 crc kubenswrapper[4934]: I1002 09:51:51.097793 4934 generic.go:334] "Generic (PLEG): container finished" podID="ea34a4e8-c842-4a23-933a-8ecd3f882d19" containerID="3b963f52f5103d064cce85dfc5d3fedb1aa2c7c534722d0be9ba5c4f5278015e" exitCode=0 Oct 02 09:51:51 crc kubenswrapper[4934]: I1002 09:51:51.097873 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6k6dn" event={"ID":"ea34a4e8-c842-4a23-933a-8ecd3f882d19","Type":"ContainerDied","Data":"3b963f52f5103d064cce85dfc5d3fedb1aa2c7c534722d0be9ba5c4f5278015e"} Oct 02 09:51:51 crc kubenswrapper[4934]: I1002 09:51:51.102346 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-78cmv" event={"ID":"fac56d4f-c941-486b-93cb-6931afbd89e0","Type":"ContainerStarted","Data":"38ea65cbd2658ea8cc3a93bb5838e6fb6c6a3d052166ed430c880aa17f88dba3"} Oct 02 09:51:51 crc kubenswrapper[4934]: I1002 09:51:51.105279 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fdrpb" event={"ID":"b8150488-b2c9-45e0-8209-9d6a73b64834","Type":"ContainerStarted","Data":"00b5632742793c3358edd24ff78e0322fc115b198ac65b62756a1569e9f480f9"} Oct 02 09:51:51 crc kubenswrapper[4934]: I1002 09:51:51.111224 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fj8zn" podStartSLOduration=2.102339749 podStartE2EDuration="55.111204969s" podCreationTimestamp="2025-10-02 09:50:56 +0000 UTC" firstStartedPulling="2025-10-02 09:50:57.569009809 +0000 UTC m=+129.321651331" lastFinishedPulling="2025-10-02 09:51:50.577875039 +0000 UTC m=+182.330516551" observedRunningTime="2025-10-02 09:51:51.110189473 +0000 UTC m=+182.862830995" watchObservedRunningTime="2025-10-02 09:51:51.111204969 +0000 UTC m=+182.863846491" Oct 02 09:51:51 crc kubenswrapper[4934]: I1002 09:51:51.178017 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-78cmv" podStartSLOduration=3.334346538 podStartE2EDuration="53.177989124s" podCreationTimestamp="2025-10-02 09:50:58 +0000 UTC" firstStartedPulling="2025-10-02 09:51:00.705103744 +0000 UTC m=+132.457745266" lastFinishedPulling="2025-10-02 09:51:50.54874631 +0000 UTC m=+182.301387852" observedRunningTime="2025-10-02 09:51:51.1556378 +0000 UTC m=+182.908279332" watchObservedRunningTime="2025-10-02 09:51:51.177989124 +0000 UTC m=+182.930630646" Oct 02 09:51:51 crc kubenswrapper[4934]: I1002 09:51:51.184555 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fdrpb" podStartSLOduration=2.285403137 podStartE2EDuration="55.184535962s" podCreationTimestamp="2025-10-02 09:50:56 +0000 UTC" firstStartedPulling="2025-10-02 09:50:57.582266425 +0000 UTC m=+129.334907947" lastFinishedPulling="2025-10-02 09:51:50.48139925 +0000 UTC m=+182.234040772" observedRunningTime="2025-10-02 09:51:51.174919305 +0000 UTC m=+182.927560827" watchObservedRunningTime="2025-10-02 09:51:51.184535962 +0000 UTC m=+182.937177484" Oct 02 09:51:52 crc kubenswrapper[4934]: I1002 09:51:52.114084 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6k6dn" event={"ID":"ea34a4e8-c842-4a23-933a-8ecd3f882d19","Type":"ContainerStarted","Data":"5a83f9d75250d2d008490d6b45e41978f650b5e62519f19429d6732dabd03e7c"} Oct 02 09:51:53 crc kubenswrapper[4934]: I1002 09:51:53.254164 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6k6dn" podStartSLOduration=4.298848958 podStartE2EDuration="58.254144235s" podCreationTimestamp="2025-10-02 09:50:55 +0000 UTC" firstStartedPulling="2025-10-02 09:50:57.575540221 +0000 UTC m=+129.328181743" lastFinishedPulling="2025-10-02 09:51:51.530835498 +0000 UTC m=+183.283477020" observedRunningTime="2025-10-02 09:51:52.138009455 +0000 UTC m=+183.890650977" watchObservedRunningTime="2025-10-02 09:51:53.254144235 +0000 UTC m=+185.006785757" Oct 02 09:51:53 crc kubenswrapper[4934]: I1002 09:51:53.255877 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lrg5q"] Oct 02 09:51:53 crc kubenswrapper[4934]: I1002 09:51:53.256135 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lrg5q" podUID="be630036-c659-47b0-a7bb-bac356b7df7a" containerName="registry-server" containerID="cri-o://6ea364178614fd8fd8bf29bca4e2adea528f0f02d7ff35f2f0057df68378b81f" gracePeriod=2 Oct 02 09:51:54 crc kubenswrapper[4934]: I1002 09:51:54.125887 4934 generic.go:334] "Generic (PLEG): container finished" podID="be630036-c659-47b0-a7bb-bac356b7df7a" containerID="6ea364178614fd8fd8bf29bca4e2adea528f0f02d7ff35f2f0057df68378b81f" exitCode=0 Oct 02 09:51:54 crc kubenswrapper[4934]: I1002 09:51:54.125975 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrg5q" event={"ID":"be630036-c659-47b0-a7bb-bac356b7df7a","Type":"ContainerDied","Data":"6ea364178614fd8fd8bf29bca4e2adea528f0f02d7ff35f2f0057df68378b81f"} Oct 02 09:51:54 crc kubenswrapper[4934]: I1002 09:51:54.557937 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lrg5q" Oct 02 09:51:54 crc kubenswrapper[4934]: I1002 09:51:54.730363 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xv2bp\" (UniqueName: \"kubernetes.io/projected/be630036-c659-47b0-a7bb-bac356b7df7a-kube-api-access-xv2bp\") pod \"be630036-c659-47b0-a7bb-bac356b7df7a\" (UID: \"be630036-c659-47b0-a7bb-bac356b7df7a\") " Oct 02 09:51:54 crc kubenswrapper[4934]: I1002 09:51:54.730549 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be630036-c659-47b0-a7bb-bac356b7df7a-utilities\") pod \"be630036-c659-47b0-a7bb-bac356b7df7a\" (UID: \"be630036-c659-47b0-a7bb-bac356b7df7a\") " Oct 02 09:51:54 crc kubenswrapper[4934]: I1002 09:51:54.730604 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be630036-c659-47b0-a7bb-bac356b7df7a-catalog-content\") pod \"be630036-c659-47b0-a7bb-bac356b7df7a\" (UID: \"be630036-c659-47b0-a7bb-bac356b7df7a\") " Oct 02 09:51:54 crc kubenswrapper[4934]: I1002 09:51:54.731649 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be630036-c659-47b0-a7bb-bac356b7df7a-utilities" (OuterVolumeSpecName: "utilities") pod "be630036-c659-47b0-a7bb-bac356b7df7a" (UID: "be630036-c659-47b0-a7bb-bac356b7df7a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:51:54 crc kubenswrapper[4934]: I1002 09:51:54.738291 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be630036-c659-47b0-a7bb-bac356b7df7a-kube-api-access-xv2bp" (OuterVolumeSpecName: "kube-api-access-xv2bp") pod "be630036-c659-47b0-a7bb-bac356b7df7a" (UID: "be630036-c659-47b0-a7bb-bac356b7df7a"). InnerVolumeSpecName "kube-api-access-xv2bp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:51:54 crc kubenswrapper[4934]: I1002 09:51:54.822254 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be630036-c659-47b0-a7bb-bac356b7df7a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "be630036-c659-47b0-a7bb-bac356b7df7a" (UID: "be630036-c659-47b0-a7bb-bac356b7df7a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:51:54 crc kubenswrapper[4934]: I1002 09:51:54.831568 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be630036-c659-47b0-a7bb-bac356b7df7a-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:51:54 crc kubenswrapper[4934]: I1002 09:51:54.831706 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be630036-c659-47b0-a7bb-bac356b7df7a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:51:54 crc kubenswrapper[4934]: I1002 09:51:54.831720 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xv2bp\" (UniqueName: \"kubernetes.io/projected/be630036-c659-47b0-a7bb-bac356b7df7a-kube-api-access-xv2bp\") on node \"crc\" DevicePath \"\"" Oct 02 09:51:55 crc kubenswrapper[4934]: I1002 09:51:55.132664 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrg5q" event={"ID":"be630036-c659-47b0-a7bb-bac356b7df7a","Type":"ContainerDied","Data":"3083ae48699cfb77960ebfffa8d17f427a14e293ae41fa0b0b97c4f9338cfadf"} Oct 02 09:51:55 crc kubenswrapper[4934]: I1002 09:51:55.132725 4934 scope.go:117] "RemoveContainer" containerID="6ea364178614fd8fd8bf29bca4e2adea528f0f02d7ff35f2f0057df68378b81f" Oct 02 09:51:55 crc kubenswrapper[4934]: I1002 09:51:55.132879 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lrg5q" Oct 02 09:51:55 crc kubenswrapper[4934]: I1002 09:51:55.155038 4934 scope.go:117] "RemoveContainer" containerID="bbbd47eb40cc85d7be63d1285bbe262a125707e1555597419fe066309816b520" Oct 02 09:51:55 crc kubenswrapper[4934]: I1002 09:51:55.157304 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lrg5q"] Oct 02 09:51:55 crc kubenswrapper[4934]: I1002 09:51:55.163949 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lrg5q"] Oct 02 09:51:55 crc kubenswrapper[4934]: I1002 09:51:55.169690 4934 scope.go:117] "RemoveContainer" containerID="3954675970c95e9b2f83cbd67321838b462d4590c2c82e624a4e51f1586973e4" Oct 02 09:51:56 crc kubenswrapper[4934]: I1002 09:51:56.272678 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6k6dn" Oct 02 09:51:56 crc kubenswrapper[4934]: I1002 09:51:56.272740 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6k6dn" Oct 02 09:51:56 crc kubenswrapper[4934]: I1002 09:51:56.311118 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6k6dn" Oct 02 09:51:56 crc kubenswrapper[4934]: I1002 09:51:56.471370 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fdrpb" Oct 02 09:51:56 crc kubenswrapper[4934]: I1002 09:51:56.471437 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fdrpb" Oct 02 09:51:56 crc kubenswrapper[4934]: I1002 09:51:56.505000 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fdrpb" Oct 02 09:51:56 crc kubenswrapper[4934]: I1002 09:51:56.691210 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fj8zn" Oct 02 09:51:56 crc kubenswrapper[4934]: I1002 09:51:56.691287 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fj8zn" Oct 02 09:51:56 crc kubenswrapper[4934]: I1002 09:51:56.730680 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fj8zn" Oct 02 09:51:56 crc kubenswrapper[4934]: I1002 09:51:56.923463 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be630036-c659-47b0-a7bb-bac356b7df7a" path="/var/lib/kubelet/pods/be630036-c659-47b0-a7bb-bac356b7df7a/volumes" Oct 02 09:51:57 crc kubenswrapper[4934]: I1002 09:51:57.147523 4934 generic.go:334] "Generic (PLEG): container finished" podID="e0220445-630a-4d62-91c9-ba260e6ba945" containerID="25c0b2eb473bb00a3d477633c77913152321322f706a428d74279646f7563e02" exitCode=0 Oct 02 09:51:57 crc kubenswrapper[4934]: I1002 09:51:57.147654 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-llmzg" event={"ID":"e0220445-630a-4d62-91c9-ba260e6ba945","Type":"ContainerDied","Data":"25c0b2eb473bb00a3d477633c77913152321322f706a428d74279646f7563e02"} Oct 02 09:51:57 crc kubenswrapper[4934]: I1002 09:51:57.194479 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6k6dn" Oct 02 09:51:57 crc kubenswrapper[4934]: I1002 09:51:57.196118 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fj8zn" Oct 02 09:51:57 crc kubenswrapper[4934]: I1002 09:51:57.198725 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fdrpb" Oct 02 09:51:58 crc kubenswrapper[4934]: I1002 09:51:58.471320 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-78cmv" Oct 02 09:51:58 crc kubenswrapper[4934]: I1002 09:51:58.472136 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-78cmv" Oct 02 09:51:58 crc kubenswrapper[4934]: I1002 09:51:58.520827 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-78cmv" Oct 02 09:51:58 crc kubenswrapper[4934]: I1002 09:51:58.969924 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mz6xf" Oct 02 09:51:59 crc kubenswrapper[4934]: I1002 09:51:59.201960 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-78cmv" Oct 02 09:51:59 crc kubenswrapper[4934]: I1002 09:51:59.459395 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fj8zn"] Oct 02 09:51:59 crc kubenswrapper[4934]: I1002 09:51:59.459931 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fj8zn" podUID="ab1258e2-b9fc-44da-8e7d-c9e70cd7897d" containerName="registry-server" containerID="cri-o://f948029f5bf66c725b94a8a7b743870b6258a1dd1ac6482c29f587702a519a61" gracePeriod=2 Oct 02 09:52:00 crc kubenswrapper[4934]: I1002 09:52:00.164974 4934 generic.go:334] "Generic (PLEG): container finished" podID="ab1258e2-b9fc-44da-8e7d-c9e70cd7897d" containerID="f948029f5bf66c725b94a8a7b743870b6258a1dd1ac6482c29f587702a519a61" exitCode=0 Oct 02 09:52:00 crc kubenswrapper[4934]: I1002 09:52:00.165014 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fj8zn" event={"ID":"ab1258e2-b9fc-44da-8e7d-c9e70cd7897d","Type":"ContainerDied","Data":"f948029f5bf66c725b94a8a7b743870b6258a1dd1ac6482c29f587702a519a61"} Oct 02 09:52:00 crc kubenswrapper[4934]: I1002 09:52:00.165326 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fj8zn" event={"ID":"ab1258e2-b9fc-44da-8e7d-c9e70cd7897d","Type":"ContainerDied","Data":"80feceba4fa69796d9846e77171252701feadf88972ac15345990b7d18b90b32"} Oct 02 09:52:00 crc kubenswrapper[4934]: I1002 09:52:00.165351 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80feceba4fa69796d9846e77171252701feadf88972ac15345990b7d18b90b32" Oct 02 09:52:00 crc kubenswrapper[4934]: I1002 09:52:00.196540 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fj8zn" Oct 02 09:52:00 crc kubenswrapper[4934]: I1002 09:52:00.302197 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab1258e2-b9fc-44da-8e7d-c9e70cd7897d-catalog-content\") pod \"ab1258e2-b9fc-44da-8e7d-c9e70cd7897d\" (UID: \"ab1258e2-b9fc-44da-8e7d-c9e70cd7897d\") " Oct 02 09:52:00 crc kubenswrapper[4934]: I1002 09:52:00.302245 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab1258e2-b9fc-44da-8e7d-c9e70cd7897d-utilities\") pod \"ab1258e2-b9fc-44da-8e7d-c9e70cd7897d\" (UID: \"ab1258e2-b9fc-44da-8e7d-c9e70cd7897d\") " Oct 02 09:52:00 crc kubenswrapper[4934]: I1002 09:52:00.302331 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wj2fk\" (UniqueName: \"kubernetes.io/projected/ab1258e2-b9fc-44da-8e7d-c9e70cd7897d-kube-api-access-wj2fk\") pod \"ab1258e2-b9fc-44da-8e7d-c9e70cd7897d\" (UID: \"ab1258e2-b9fc-44da-8e7d-c9e70cd7897d\") " Oct 02 09:52:00 crc kubenswrapper[4934]: I1002 09:52:00.303223 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab1258e2-b9fc-44da-8e7d-c9e70cd7897d-utilities" (OuterVolumeSpecName: "utilities") pod "ab1258e2-b9fc-44da-8e7d-c9e70cd7897d" (UID: "ab1258e2-b9fc-44da-8e7d-c9e70cd7897d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:52:00 crc kubenswrapper[4934]: I1002 09:52:00.308807 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab1258e2-b9fc-44da-8e7d-c9e70cd7897d-kube-api-access-wj2fk" (OuterVolumeSpecName: "kube-api-access-wj2fk") pod "ab1258e2-b9fc-44da-8e7d-c9e70cd7897d" (UID: "ab1258e2-b9fc-44da-8e7d-c9e70cd7897d"). InnerVolumeSpecName "kube-api-access-wj2fk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:52:00 crc kubenswrapper[4934]: I1002 09:52:00.349222 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab1258e2-b9fc-44da-8e7d-c9e70cd7897d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ab1258e2-b9fc-44da-8e7d-c9e70cd7897d" (UID: "ab1258e2-b9fc-44da-8e7d-c9e70cd7897d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:52:00 crc kubenswrapper[4934]: I1002 09:52:00.403511 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wj2fk\" (UniqueName: \"kubernetes.io/projected/ab1258e2-b9fc-44da-8e7d-c9e70cd7897d-kube-api-access-wj2fk\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:00 crc kubenswrapper[4934]: I1002 09:52:00.403551 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab1258e2-b9fc-44da-8e7d-c9e70cd7897d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:00 crc kubenswrapper[4934]: I1002 09:52:00.403563 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab1258e2-b9fc-44da-8e7d-c9e70cd7897d-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:01 crc kubenswrapper[4934]: I1002 09:52:01.180331 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-llmzg" event={"ID":"e0220445-630a-4d62-91c9-ba260e6ba945","Type":"ContainerStarted","Data":"eda1437a5f9528c10a8c8e5b0efa4c85266c338e826b4eb346ca04f5499743e4"} Oct 02 09:52:01 crc kubenswrapper[4934]: I1002 09:52:01.180402 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fj8zn" Oct 02 09:52:01 crc kubenswrapper[4934]: I1002 09:52:01.211904 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fj8zn"] Oct 02 09:52:01 crc kubenswrapper[4934]: I1002 09:52:01.219325 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fj8zn"] Oct 02 09:52:01 crc kubenswrapper[4934]: I1002 09:52:01.655545 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-llmzg" podStartSLOduration=3.43155211 podStartE2EDuration="1m5.655530183s" podCreationTimestamp="2025-10-02 09:50:56 +0000 UTC" firstStartedPulling="2025-10-02 09:50:57.577549016 +0000 UTC m=+129.330190528" lastFinishedPulling="2025-10-02 09:51:59.801527069 +0000 UTC m=+191.554168601" observedRunningTime="2025-10-02 09:52:01.234075558 +0000 UTC m=+192.986717110" watchObservedRunningTime="2025-10-02 09:52:01.655530183 +0000 UTC m=+193.408171705" Oct 02 09:52:01 crc kubenswrapper[4934]: I1002 09:52:01.658527 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mz6xf"] Oct 02 09:52:01 crc kubenswrapper[4934]: I1002 09:52:01.658800 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mz6xf" podUID="a7c3f392-8c78-4a39-ac9f-d16cb2832698" containerName="registry-server" containerID="cri-o://5713b8fe75a793097de0412c1ed9fdf782e19786308d56da8fa45a3a03ed843e" gracePeriod=2 Oct 02 09:52:02 crc kubenswrapper[4934]: I1002 09:52:02.198291 4934 generic.go:334] "Generic (PLEG): container finished" podID="a7c3f392-8c78-4a39-ac9f-d16cb2832698" containerID="5713b8fe75a793097de0412c1ed9fdf782e19786308d56da8fa45a3a03ed843e" exitCode=0 Oct 02 09:52:02 crc kubenswrapper[4934]: I1002 09:52:02.198331 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mz6xf" event={"ID":"a7c3f392-8c78-4a39-ac9f-d16cb2832698","Type":"ContainerDied","Data":"5713b8fe75a793097de0412c1ed9fdf782e19786308d56da8fa45a3a03ed843e"} Oct 02 09:52:02 crc kubenswrapper[4934]: I1002 09:52:02.635768 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mz6xf" Oct 02 09:52:02 crc kubenswrapper[4934]: I1002 09:52:02.743064 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7c3f392-8c78-4a39-ac9f-d16cb2832698-catalog-content\") pod \"a7c3f392-8c78-4a39-ac9f-d16cb2832698\" (UID: \"a7c3f392-8c78-4a39-ac9f-d16cb2832698\") " Oct 02 09:52:02 crc kubenswrapper[4934]: I1002 09:52:02.743149 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkgwt\" (UniqueName: \"kubernetes.io/projected/a7c3f392-8c78-4a39-ac9f-d16cb2832698-kube-api-access-jkgwt\") pod \"a7c3f392-8c78-4a39-ac9f-d16cb2832698\" (UID: \"a7c3f392-8c78-4a39-ac9f-d16cb2832698\") " Oct 02 09:52:02 crc kubenswrapper[4934]: I1002 09:52:02.743327 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7c3f392-8c78-4a39-ac9f-d16cb2832698-utilities\") pod \"a7c3f392-8c78-4a39-ac9f-d16cb2832698\" (UID: \"a7c3f392-8c78-4a39-ac9f-d16cb2832698\") " Oct 02 09:52:02 crc kubenswrapper[4934]: I1002 09:52:02.744157 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7c3f392-8c78-4a39-ac9f-d16cb2832698-utilities" (OuterVolumeSpecName: "utilities") pod "a7c3f392-8c78-4a39-ac9f-d16cb2832698" (UID: "a7c3f392-8c78-4a39-ac9f-d16cb2832698"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:52:02 crc kubenswrapper[4934]: I1002 09:52:02.744393 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7c3f392-8c78-4a39-ac9f-d16cb2832698-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:02 crc kubenswrapper[4934]: I1002 09:52:02.748853 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7c3f392-8c78-4a39-ac9f-d16cb2832698-kube-api-access-jkgwt" (OuterVolumeSpecName: "kube-api-access-jkgwt") pod "a7c3f392-8c78-4a39-ac9f-d16cb2832698" (UID: "a7c3f392-8c78-4a39-ac9f-d16cb2832698"). InnerVolumeSpecName "kube-api-access-jkgwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:52:02 crc kubenswrapper[4934]: I1002 09:52:02.773353 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7c3f392-8c78-4a39-ac9f-d16cb2832698-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a7c3f392-8c78-4a39-ac9f-d16cb2832698" (UID: "a7c3f392-8c78-4a39-ac9f-d16cb2832698"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:52:02 crc kubenswrapper[4934]: I1002 09:52:02.846243 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7c3f392-8c78-4a39-ac9f-d16cb2832698-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:02 crc kubenswrapper[4934]: I1002 09:52:02.846292 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkgwt\" (UniqueName: \"kubernetes.io/projected/a7c3f392-8c78-4a39-ac9f-d16cb2832698-kube-api-access-jkgwt\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:02 crc kubenswrapper[4934]: I1002 09:52:02.930802 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab1258e2-b9fc-44da-8e7d-c9e70cd7897d" path="/var/lib/kubelet/pods/ab1258e2-b9fc-44da-8e7d-c9e70cd7897d/volumes" Oct 02 09:52:03 crc kubenswrapper[4934]: I1002 09:52:03.207192 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mz6xf" event={"ID":"a7c3f392-8c78-4a39-ac9f-d16cb2832698","Type":"ContainerDied","Data":"ee165fb3416e086e6dd7dfca2c37aad15fc78d246964d0bfbf1ef8d80c961445"} Oct 02 09:52:03 crc kubenswrapper[4934]: I1002 09:52:03.207240 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mz6xf" Oct 02 09:52:03 crc kubenswrapper[4934]: I1002 09:52:03.207257 4934 scope.go:117] "RemoveContainer" containerID="5713b8fe75a793097de0412c1ed9fdf782e19786308d56da8fa45a3a03ed843e" Oct 02 09:52:03 crc kubenswrapper[4934]: I1002 09:52:03.224741 4934 scope.go:117] "RemoveContainer" containerID="51ec98110649f0ef94d550060bff4fe3093ec184bfb2dd4d11a76f1252d62763" Oct 02 09:52:03 crc kubenswrapper[4934]: I1002 09:52:03.250387 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mz6xf"] Oct 02 09:52:03 crc kubenswrapper[4934]: I1002 09:52:03.254489 4934 scope.go:117] "RemoveContainer" containerID="24bf0e3bafa23c98917c42b27fc71fffee12d56d5afd06458a775891755e38ac" Oct 02 09:52:03 crc kubenswrapper[4934]: I1002 09:52:03.257715 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mz6xf"] Oct 02 09:52:04 crc kubenswrapper[4934]: I1002 09:52:04.919678 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7c3f392-8c78-4a39-ac9f-d16cb2832698" path="/var/lib/kubelet/pods/a7c3f392-8c78-4a39-ac9f-d16cb2832698/volumes" Oct 02 09:52:06 crc kubenswrapper[4934]: I1002 09:52:06.895639 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-llmzg" Oct 02 09:52:06 crc kubenswrapper[4934]: I1002 09:52:06.895720 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-llmzg" Oct 02 09:52:06 crc kubenswrapper[4934]: I1002 09:52:06.930542 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-llmzg" Oct 02 09:52:07 crc kubenswrapper[4934]: I1002 09:52:07.288551 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-llmzg" Oct 02 09:52:07 crc kubenswrapper[4934]: I1002 09:52:07.856687 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-llmzg"] Oct 02 09:52:08 crc kubenswrapper[4934]: I1002 09:52:08.440128 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:52:08 crc kubenswrapper[4934]: I1002 09:52:08.440194 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:52:09 crc kubenswrapper[4934]: I1002 09:52:09.126775 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4pp8p"] Oct 02 09:52:09 crc kubenswrapper[4934]: I1002 09:52:09.235810 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-llmzg" podUID="e0220445-630a-4d62-91c9-ba260e6ba945" containerName="registry-server" containerID="cri-o://eda1437a5f9528c10a8c8e5b0efa4c85266c338e826b4eb346ca04f5499743e4" gracePeriod=2 Oct 02 09:52:09 crc kubenswrapper[4934]: I1002 09:52:09.591205 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-llmzg" Oct 02 09:52:09 crc kubenswrapper[4934]: I1002 09:52:09.750795 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jsjc\" (UniqueName: \"kubernetes.io/projected/e0220445-630a-4d62-91c9-ba260e6ba945-kube-api-access-2jsjc\") pod \"e0220445-630a-4d62-91c9-ba260e6ba945\" (UID: \"e0220445-630a-4d62-91c9-ba260e6ba945\") " Oct 02 09:52:09 crc kubenswrapper[4934]: I1002 09:52:09.751230 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0220445-630a-4d62-91c9-ba260e6ba945-utilities\") pod \"e0220445-630a-4d62-91c9-ba260e6ba945\" (UID: \"e0220445-630a-4d62-91c9-ba260e6ba945\") " Oct 02 09:52:09 crc kubenswrapper[4934]: I1002 09:52:09.751287 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0220445-630a-4d62-91c9-ba260e6ba945-catalog-content\") pod \"e0220445-630a-4d62-91c9-ba260e6ba945\" (UID: \"e0220445-630a-4d62-91c9-ba260e6ba945\") " Oct 02 09:52:09 crc kubenswrapper[4934]: I1002 09:52:09.751899 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0220445-630a-4d62-91c9-ba260e6ba945-utilities" (OuterVolumeSpecName: "utilities") pod "e0220445-630a-4d62-91c9-ba260e6ba945" (UID: "e0220445-630a-4d62-91c9-ba260e6ba945"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:52:09 crc kubenswrapper[4934]: I1002 09:52:09.762475 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0220445-630a-4d62-91c9-ba260e6ba945-kube-api-access-2jsjc" (OuterVolumeSpecName: "kube-api-access-2jsjc") pod "e0220445-630a-4d62-91c9-ba260e6ba945" (UID: "e0220445-630a-4d62-91c9-ba260e6ba945"). InnerVolumeSpecName "kube-api-access-2jsjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:52:09 crc kubenswrapper[4934]: I1002 09:52:09.805084 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0220445-630a-4d62-91c9-ba260e6ba945-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e0220445-630a-4d62-91c9-ba260e6ba945" (UID: "e0220445-630a-4d62-91c9-ba260e6ba945"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:52:09 crc kubenswrapper[4934]: I1002 09:52:09.852097 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0220445-630a-4d62-91c9-ba260e6ba945-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:09 crc kubenswrapper[4934]: I1002 09:52:09.852123 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0220445-630a-4d62-91c9-ba260e6ba945-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:09 crc kubenswrapper[4934]: I1002 09:52:09.852133 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jsjc\" (UniqueName: \"kubernetes.io/projected/e0220445-630a-4d62-91c9-ba260e6ba945-kube-api-access-2jsjc\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:10 crc kubenswrapper[4934]: I1002 09:52:10.242211 4934 generic.go:334] "Generic (PLEG): container finished" podID="e0220445-630a-4d62-91c9-ba260e6ba945" containerID="eda1437a5f9528c10a8c8e5b0efa4c85266c338e826b4eb346ca04f5499743e4" exitCode=0 Oct 02 09:52:10 crc kubenswrapper[4934]: I1002 09:52:10.242268 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-llmzg" event={"ID":"e0220445-630a-4d62-91c9-ba260e6ba945","Type":"ContainerDied","Data":"eda1437a5f9528c10a8c8e5b0efa4c85266c338e826b4eb346ca04f5499743e4"} Oct 02 09:52:10 crc kubenswrapper[4934]: I1002 09:52:10.242293 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-llmzg" event={"ID":"e0220445-630a-4d62-91c9-ba260e6ba945","Type":"ContainerDied","Data":"24c0f821815b6b1821c6c9866e3b95f9178b900d9b4165882b0550fbb52d0d4a"} Oct 02 09:52:10 crc kubenswrapper[4934]: I1002 09:52:10.242326 4934 scope.go:117] "RemoveContainer" containerID="eda1437a5f9528c10a8c8e5b0efa4c85266c338e826b4eb346ca04f5499743e4" Oct 02 09:52:10 crc kubenswrapper[4934]: I1002 09:52:10.242382 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-llmzg" Oct 02 09:52:10 crc kubenswrapper[4934]: I1002 09:52:10.256637 4934 scope.go:117] "RemoveContainer" containerID="25c0b2eb473bb00a3d477633c77913152321322f706a428d74279646f7563e02" Oct 02 09:52:10 crc kubenswrapper[4934]: I1002 09:52:10.273537 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-llmzg"] Oct 02 09:52:10 crc kubenswrapper[4934]: I1002 09:52:10.275439 4934 scope.go:117] "RemoveContainer" containerID="e8937a6c4cedc22c9713537c06d00a9b28cba540f381db5bc4ef5ce0e74a5313" Oct 02 09:52:10 crc kubenswrapper[4934]: I1002 09:52:10.276760 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-llmzg"] Oct 02 09:52:10 crc kubenswrapper[4934]: I1002 09:52:10.299463 4934 scope.go:117] "RemoveContainer" containerID="eda1437a5f9528c10a8c8e5b0efa4c85266c338e826b4eb346ca04f5499743e4" Oct 02 09:52:10 crc kubenswrapper[4934]: E1002 09:52:10.299916 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eda1437a5f9528c10a8c8e5b0efa4c85266c338e826b4eb346ca04f5499743e4\": container with ID starting with eda1437a5f9528c10a8c8e5b0efa4c85266c338e826b4eb346ca04f5499743e4 not found: ID does not exist" containerID="eda1437a5f9528c10a8c8e5b0efa4c85266c338e826b4eb346ca04f5499743e4" Oct 02 09:52:10 crc kubenswrapper[4934]: I1002 09:52:10.299973 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eda1437a5f9528c10a8c8e5b0efa4c85266c338e826b4eb346ca04f5499743e4"} err="failed to get container status \"eda1437a5f9528c10a8c8e5b0efa4c85266c338e826b4eb346ca04f5499743e4\": rpc error: code = NotFound desc = could not find container \"eda1437a5f9528c10a8c8e5b0efa4c85266c338e826b4eb346ca04f5499743e4\": container with ID starting with eda1437a5f9528c10a8c8e5b0efa4c85266c338e826b4eb346ca04f5499743e4 not found: ID does not exist" Oct 02 09:52:10 crc kubenswrapper[4934]: I1002 09:52:10.300028 4934 scope.go:117] "RemoveContainer" containerID="25c0b2eb473bb00a3d477633c77913152321322f706a428d74279646f7563e02" Oct 02 09:52:10 crc kubenswrapper[4934]: E1002 09:52:10.304827 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25c0b2eb473bb00a3d477633c77913152321322f706a428d74279646f7563e02\": container with ID starting with 25c0b2eb473bb00a3d477633c77913152321322f706a428d74279646f7563e02 not found: ID does not exist" containerID="25c0b2eb473bb00a3d477633c77913152321322f706a428d74279646f7563e02" Oct 02 09:52:10 crc kubenswrapper[4934]: I1002 09:52:10.304902 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25c0b2eb473bb00a3d477633c77913152321322f706a428d74279646f7563e02"} err="failed to get container status \"25c0b2eb473bb00a3d477633c77913152321322f706a428d74279646f7563e02\": rpc error: code = NotFound desc = could not find container \"25c0b2eb473bb00a3d477633c77913152321322f706a428d74279646f7563e02\": container with ID starting with 25c0b2eb473bb00a3d477633c77913152321322f706a428d74279646f7563e02 not found: ID does not exist" Oct 02 09:52:10 crc kubenswrapper[4934]: I1002 09:52:10.304962 4934 scope.go:117] "RemoveContainer" containerID="e8937a6c4cedc22c9713537c06d00a9b28cba540f381db5bc4ef5ce0e74a5313" Oct 02 09:52:10 crc kubenswrapper[4934]: E1002 09:52:10.305447 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8937a6c4cedc22c9713537c06d00a9b28cba540f381db5bc4ef5ce0e74a5313\": container with ID starting with e8937a6c4cedc22c9713537c06d00a9b28cba540f381db5bc4ef5ce0e74a5313 not found: ID does not exist" containerID="e8937a6c4cedc22c9713537c06d00a9b28cba540f381db5bc4ef5ce0e74a5313" Oct 02 09:52:10 crc kubenswrapper[4934]: I1002 09:52:10.305484 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8937a6c4cedc22c9713537c06d00a9b28cba540f381db5bc4ef5ce0e74a5313"} err="failed to get container status \"e8937a6c4cedc22c9713537c06d00a9b28cba540f381db5bc4ef5ce0e74a5313\": rpc error: code = NotFound desc = could not find container \"e8937a6c4cedc22c9713537c06d00a9b28cba540f381db5bc4ef5ce0e74a5313\": container with ID starting with e8937a6c4cedc22c9713537c06d00a9b28cba540f381db5bc4ef5ce0e74a5313 not found: ID does not exist" Oct 02 09:52:10 crc kubenswrapper[4934]: I1002 09:52:10.920097 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0220445-630a-4d62-91c9-ba260e6ba945" path="/var/lib/kubelet/pods/e0220445-630a-4d62-91c9-ba260e6ba945/volumes" Oct 02 09:52:16 crc kubenswrapper[4934]: I1002 09:52:16.241604 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.153723 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" podUID="962afee9-b32a-40cc-b6ca-c8e1cb95c13f" containerName="oauth-openshift" containerID="cri-o://7faa43dce95ff9b51790c77d99013b25fa2bb327e13eae8ad536c5870acd730c" gracePeriod=15 Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.392969 4934 generic.go:334] "Generic (PLEG): container finished" podID="962afee9-b32a-40cc-b6ca-c8e1cb95c13f" containerID="7faa43dce95ff9b51790c77d99013b25fa2bb327e13eae8ad536c5870acd730c" exitCode=0 Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.393505 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" event={"ID":"962afee9-b32a-40cc-b6ca-c8e1cb95c13f","Type":"ContainerDied","Data":"7faa43dce95ff9b51790c77d99013b25fa2bb327e13eae8ad536c5870acd730c"} Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.618298 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.675677 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-55889b984c-nrvll"] Oct 02 09:52:34 crc kubenswrapper[4934]: E1002 09:52:34.676089 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7c3f392-8c78-4a39-ac9f-d16cb2832698" containerName="extract-content" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.676125 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7c3f392-8c78-4a39-ac9f-d16cb2832698" containerName="extract-content" Oct 02 09:52:34 crc kubenswrapper[4934]: E1002 09:52:34.676155 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be630036-c659-47b0-a7bb-bac356b7df7a" containerName="extract-utilities" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.676173 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="be630036-c659-47b0-a7bb-bac356b7df7a" containerName="extract-utilities" Oct 02 09:52:34 crc kubenswrapper[4934]: E1002 09:52:34.676208 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be630036-c659-47b0-a7bb-bac356b7df7a" containerName="registry-server" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.676226 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="be630036-c659-47b0-a7bb-bac356b7df7a" containerName="registry-server" Oct 02 09:52:34 crc kubenswrapper[4934]: E1002 09:52:34.676253 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be630036-c659-47b0-a7bb-bac356b7df7a" containerName="extract-content" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.676270 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="be630036-c659-47b0-a7bb-bac356b7df7a" containerName="extract-content" Oct 02 09:52:34 crc kubenswrapper[4934]: E1002 09:52:34.676289 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7c3f392-8c78-4a39-ac9f-d16cb2832698" containerName="registry-server" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.676306 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7c3f392-8c78-4a39-ac9f-d16cb2832698" containerName="registry-server" Oct 02 09:52:34 crc kubenswrapper[4934]: E1002 09:52:34.676333 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0c99a28-1f5b-4a44-a03c-e9c3b3254154" containerName="pruner" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.676350 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0c99a28-1f5b-4a44-a03c-e9c3b3254154" containerName="pruner" Oct 02 09:52:34 crc kubenswrapper[4934]: E1002 09:52:34.676368 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab1258e2-b9fc-44da-8e7d-c9e70cd7897d" containerName="registry-server" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.676387 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab1258e2-b9fc-44da-8e7d-c9e70cd7897d" containerName="registry-server" Oct 02 09:52:34 crc kubenswrapper[4934]: E1002 09:52:34.676407 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0220445-630a-4d62-91c9-ba260e6ba945" containerName="extract-utilities" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.676422 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0220445-630a-4d62-91c9-ba260e6ba945" containerName="extract-utilities" Oct 02 09:52:34 crc kubenswrapper[4934]: E1002 09:52:34.676449 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7c3f392-8c78-4a39-ac9f-d16cb2832698" containerName="extract-utilities" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.676465 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7c3f392-8c78-4a39-ac9f-d16cb2832698" containerName="extract-utilities" Oct 02 09:52:34 crc kubenswrapper[4934]: E1002 09:52:34.676489 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0220445-630a-4d62-91c9-ba260e6ba945" containerName="extract-content" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.676506 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0220445-630a-4d62-91c9-ba260e6ba945" containerName="extract-content" Oct 02 09:52:34 crc kubenswrapper[4934]: E1002 09:52:34.676525 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0220445-630a-4d62-91c9-ba260e6ba945" containerName="registry-server" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.676541 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0220445-630a-4d62-91c9-ba260e6ba945" containerName="registry-server" Oct 02 09:52:34 crc kubenswrapper[4934]: E1002 09:52:34.676604 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab1258e2-b9fc-44da-8e7d-c9e70cd7897d" containerName="extract-content" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.676625 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab1258e2-b9fc-44da-8e7d-c9e70cd7897d" containerName="extract-content" Oct 02 09:52:34 crc kubenswrapper[4934]: E1002 09:52:34.676651 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab1258e2-b9fc-44da-8e7d-c9e70cd7897d" containerName="extract-utilities" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.676667 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab1258e2-b9fc-44da-8e7d-c9e70cd7897d" containerName="extract-utilities" Oct 02 09:52:34 crc kubenswrapper[4934]: E1002 09:52:34.676688 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="962afee9-b32a-40cc-b6ca-c8e1cb95c13f" containerName="oauth-openshift" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.676705 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="962afee9-b32a-40cc-b6ca-c8e1cb95c13f" containerName="oauth-openshift" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.676929 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0c99a28-1f5b-4a44-a03c-e9c3b3254154" containerName="pruner" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.676953 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="be630036-c659-47b0-a7bb-bac356b7df7a" containerName="registry-server" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.676989 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7c3f392-8c78-4a39-ac9f-d16cb2832698" containerName="registry-server" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.677011 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0220445-630a-4d62-91c9-ba260e6ba945" containerName="registry-server" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.677035 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab1258e2-b9fc-44da-8e7d-c9e70cd7897d" containerName="registry-server" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.677252 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="962afee9-b32a-40cc-b6ca-c8e1cb95c13f" containerName="oauth-openshift" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.678138 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.684922 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-55889b984c-nrvll"] Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.753793 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-session\") pod \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.754050 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-service-ca\") pod \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.754107 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-audit-dir\") pod \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.754149 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-user-idp-0-file-data\") pod \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.754195 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-ocp-branding-template\") pod \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.754240 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-audit-policies\") pod \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.754251 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "962afee9-b32a-40cc-b6ca-c8e1cb95c13f" (UID: "962afee9-b32a-40cc-b6ca-c8e1cb95c13f"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.754278 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-router-certs\") pod \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.754439 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-serving-cert\") pod \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.754515 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-cliconfig\") pod \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.754623 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rc58m\" (UniqueName: \"kubernetes.io/projected/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-kube-api-access-rc58m\") pod \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.754663 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-trusted-ca-bundle\") pod \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.754719 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-user-template-error\") pod \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.754770 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-user-template-login\") pod \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.754967 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-user-template-provider-selection\") pod \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\" (UID: \"962afee9-b32a-40cc-b6ca-c8e1cb95c13f\") " Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.754981 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "962afee9-b32a-40cc-b6ca-c8e1cb95c13f" (UID: "962afee9-b32a-40cc-b6ca-c8e1cb95c13f"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.754992 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "962afee9-b32a-40cc-b6ca-c8e1cb95c13f" (UID: "962afee9-b32a-40cc-b6ca-c8e1cb95c13f"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.755189 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-audit-dir\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.755692 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "962afee9-b32a-40cc-b6ca-c8e1cb95c13f" (UID: "962afee9-b32a-40cc-b6ca-c8e1cb95c13f"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.756233 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "962afee9-b32a-40cc-b6ca-c8e1cb95c13f" (UID: "962afee9-b32a-40cc-b6ca-c8e1cb95c13f"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.756478 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.756674 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.756814 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.756944 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-user-template-error\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.757002 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5h4fb\" (UniqueName: \"kubernetes.io/projected/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-kube-api-access-5h4fb\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.757054 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.757082 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.757107 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.757124 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-system-router-certs\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.757149 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-user-template-login\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.757174 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-system-session\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.757269 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-audit-policies\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.757342 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-system-service-ca\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.757465 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.757482 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.757498 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.757512 4934 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.757524 4934 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.760114 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "962afee9-b32a-40cc-b6ca-c8e1cb95c13f" (UID: "962afee9-b32a-40cc-b6ca-c8e1cb95c13f"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.760445 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "962afee9-b32a-40cc-b6ca-c8e1cb95c13f" (UID: "962afee9-b32a-40cc-b6ca-c8e1cb95c13f"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.760946 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "962afee9-b32a-40cc-b6ca-c8e1cb95c13f" (UID: "962afee9-b32a-40cc-b6ca-c8e1cb95c13f"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.762845 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "962afee9-b32a-40cc-b6ca-c8e1cb95c13f" (UID: "962afee9-b32a-40cc-b6ca-c8e1cb95c13f"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.762872 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-kube-api-access-rc58m" (OuterVolumeSpecName: "kube-api-access-rc58m") pod "962afee9-b32a-40cc-b6ca-c8e1cb95c13f" (UID: "962afee9-b32a-40cc-b6ca-c8e1cb95c13f"). InnerVolumeSpecName "kube-api-access-rc58m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.763197 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "962afee9-b32a-40cc-b6ca-c8e1cb95c13f" (UID: "962afee9-b32a-40cc-b6ca-c8e1cb95c13f"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.763569 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "962afee9-b32a-40cc-b6ca-c8e1cb95c13f" (UID: "962afee9-b32a-40cc-b6ca-c8e1cb95c13f"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.763869 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "962afee9-b32a-40cc-b6ca-c8e1cb95c13f" (UID: "962afee9-b32a-40cc-b6ca-c8e1cb95c13f"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.764207 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "962afee9-b32a-40cc-b6ca-c8e1cb95c13f" (UID: "962afee9-b32a-40cc-b6ca-c8e1cb95c13f"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.858127 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.858175 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-system-router-certs\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.858204 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-user-template-login\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.858229 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-system-session\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.858250 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-audit-policies\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.858275 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-system-service-ca\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.858311 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-audit-dir\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.858335 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.858358 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.858404 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.858435 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-user-template-error\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.858455 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5h4fb\" (UniqueName: \"kubernetes.io/projected/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-kube-api-access-5h4fb\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.858481 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.858506 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.858603 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.858616 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.858627 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.858638 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.858647 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rc58m\" (UniqueName: \"kubernetes.io/projected/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-kube-api-access-rc58m\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.858658 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.858668 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.858679 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.858689 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/962afee9-b32a-40cc-b6ca-c8e1cb95c13f-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.859551 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.859640 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-audit-policies\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.859864 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-system-service-ca\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.859988 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.860071 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-audit-dir\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.861928 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.862033 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-user-template-login\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.862052 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-system-session\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.862270 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.862289 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-system-router-certs\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.863368 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.864108 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-user-template-error\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.864170 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.877279 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5h4fb\" (UniqueName: \"kubernetes.io/projected/718c2a19-fed0-4e8a-a4e4-19a714feb7f6-kube-api-access-5h4fb\") pod \"oauth-openshift-55889b984c-nrvll\" (UID: \"718c2a19-fed0-4e8a-a4e4-19a714feb7f6\") " pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:34 crc kubenswrapper[4934]: I1002 09:52:34.998039 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:35 crc kubenswrapper[4934]: I1002 09:52:35.400725 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" event={"ID":"962afee9-b32a-40cc-b6ca-c8e1cb95c13f","Type":"ContainerDied","Data":"60c397c9dc0b691e29e2158d3fcc21e662f8aa98a09665105e2da4702903a607"} Oct 02 09:52:35 crc kubenswrapper[4934]: I1002 09:52:35.400995 4934 scope.go:117] "RemoveContainer" containerID="7faa43dce95ff9b51790c77d99013b25fa2bb327e13eae8ad536c5870acd730c" Oct 02 09:52:35 crc kubenswrapper[4934]: I1002 09:52:35.400803 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-4pp8p" Oct 02 09:52:35 crc kubenswrapper[4934]: I1002 09:52:35.426557 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4pp8p"] Oct 02 09:52:35 crc kubenswrapper[4934]: I1002 09:52:35.429049 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4pp8p"] Oct 02 09:52:35 crc kubenswrapper[4934]: I1002 09:52:35.431416 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-55889b984c-nrvll"] Oct 02 09:52:36 crc kubenswrapper[4934]: I1002 09:52:36.408743 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" event={"ID":"718c2a19-fed0-4e8a-a4e4-19a714feb7f6","Type":"ContainerStarted","Data":"d8263c1b6a51feaeed9b7a8016f6f07ea9417c7ea7b0a5c3e3edba544e883fc3"} Oct 02 09:52:36 crc kubenswrapper[4934]: I1002 09:52:36.409045 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" event={"ID":"718c2a19-fed0-4e8a-a4e4-19a714feb7f6","Type":"ContainerStarted","Data":"197649c6266e04719b58a1d724ceb2f149a1c34fd75f41e1fe64d24a80e269bd"} Oct 02 09:52:36 crc kubenswrapper[4934]: I1002 09:52:36.409069 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:36 crc kubenswrapper[4934]: I1002 09:52:36.413712 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" Oct 02 09:52:36 crc kubenswrapper[4934]: I1002 09:52:36.448807 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-55889b984c-nrvll" podStartSLOduration=27.448790226 podStartE2EDuration="27.448790226s" podCreationTimestamp="2025-10-02 09:52:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:52:36.428621756 +0000 UTC m=+228.181263288" watchObservedRunningTime="2025-10-02 09:52:36.448790226 +0000 UTC m=+228.201431748" Oct 02 09:52:36 crc kubenswrapper[4934]: I1002 09:52:36.920992 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="962afee9-b32a-40cc-b6ca-c8e1cb95c13f" path="/var/lib/kubelet/pods/962afee9-b32a-40cc-b6ca-c8e1cb95c13f/volumes" Oct 02 09:52:38 crc kubenswrapper[4934]: I1002 09:52:38.440087 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:52:38 crc kubenswrapper[4934]: I1002 09:52:38.440522 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:52:38 crc kubenswrapper[4934]: I1002 09:52:38.440616 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 09:52:38 crc kubenswrapper[4934]: I1002 09:52:38.441386 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af"} pod="openshift-machine-config-operator/machine-config-daemon-djh5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:52:38 crc kubenswrapper[4934]: I1002 09:52:38.441473 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" containerID="cri-o://42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af" gracePeriod=600 Oct 02 09:52:39 crc kubenswrapper[4934]: I1002 09:52:39.426809 4934 generic.go:334] "Generic (PLEG): container finished" podID="71db06ef-05b0-4f58-b251-b27117a8500a" containerID="42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af" exitCode=0 Oct 02 09:52:39 crc kubenswrapper[4934]: I1002 09:52:39.426910 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerDied","Data":"42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af"} Oct 02 09:52:39 crc kubenswrapper[4934]: I1002 09:52:39.427147 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerStarted","Data":"0b2fff80ffe94f1c29ff579444a6c9ad1f61b9fbabc6b5775d2e9293c428ca14"} Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.362498 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fdrpb"] Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.363451 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fdrpb" podUID="b8150488-b2c9-45e0-8209-9d6a73b64834" containerName="registry-server" containerID="cri-o://00b5632742793c3358edd24ff78e0322fc115b198ac65b62756a1569e9f480f9" gracePeriod=30 Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.374702 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6k6dn"] Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.375224 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6k6dn" podUID="ea34a4e8-c842-4a23-933a-8ecd3f882d19" containerName="registry-server" containerID="cri-o://5a83f9d75250d2d008490d6b45e41978f650b5e62519f19429d6732dabd03e7c" gracePeriod=30 Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.387890 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-66hrr"] Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.388142 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-66hrr" podUID="6d137db5-66a8-4066-b8d2-e88adcaaf12c" containerName="marketplace-operator" containerID="cri-o://cd49a4c22830bb661a44e96674b3557b609b43ff491066ec8fdf67883616f50d" gracePeriod=30 Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.391487 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-78cmv"] Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.391809 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-78cmv" podUID="fac56d4f-c941-486b-93cb-6931afbd89e0" containerName="registry-server" containerID="cri-o://38ea65cbd2658ea8cc3a93bb5838e6fb6c6a3d052166ed430c880aa17f88dba3" gracePeriod=30 Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.399063 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rhn6c"] Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.400696 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rhn6c" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.410729 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vn2xd"] Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.411025 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vn2xd" podUID="acf903d9-ba78-488a-b55e-6ecd290fe774" containerName="registry-server" containerID="cri-o://760adf804fc502d51d5b18c27dad88d647e7fe1ae60a645d8caa715dc86f4341" gracePeriod=30 Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.418796 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rhn6c"] Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.486545 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4f19193f-e9fe-4de1-a515-be021e499af8-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rhn6c\" (UID: \"4f19193f-e9fe-4de1-a515-be021e499af8\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhn6c" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.486613 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4f19193f-e9fe-4de1-a515-be021e499af8-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rhn6c\" (UID: \"4f19193f-e9fe-4de1-a515-be021e499af8\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhn6c" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.486771 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q8p8\" (UniqueName: \"kubernetes.io/projected/4f19193f-e9fe-4de1-a515-be021e499af8-kube-api-access-8q8p8\") pod \"marketplace-operator-79b997595-rhn6c\" (UID: \"4f19193f-e9fe-4de1-a515-be021e499af8\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhn6c" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.492771 4934 generic.go:334] "Generic (PLEG): container finished" podID="b8150488-b2c9-45e0-8209-9d6a73b64834" containerID="00b5632742793c3358edd24ff78e0322fc115b198ac65b62756a1569e9f480f9" exitCode=0 Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.492844 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fdrpb" event={"ID":"b8150488-b2c9-45e0-8209-9d6a73b64834","Type":"ContainerDied","Data":"00b5632742793c3358edd24ff78e0322fc115b198ac65b62756a1569e9f480f9"} Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.496288 4934 generic.go:334] "Generic (PLEG): container finished" podID="ea34a4e8-c842-4a23-933a-8ecd3f882d19" containerID="5a83f9d75250d2d008490d6b45e41978f650b5e62519f19429d6732dabd03e7c" exitCode=0 Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.496335 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6k6dn" event={"ID":"ea34a4e8-c842-4a23-933a-8ecd3f882d19","Type":"ContainerDied","Data":"5a83f9d75250d2d008490d6b45e41978f650b5e62519f19429d6732dabd03e7c"} Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.588006 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4f19193f-e9fe-4de1-a515-be021e499af8-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rhn6c\" (UID: \"4f19193f-e9fe-4de1-a515-be021e499af8\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhn6c" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.588065 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4f19193f-e9fe-4de1-a515-be021e499af8-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rhn6c\" (UID: \"4f19193f-e9fe-4de1-a515-be021e499af8\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhn6c" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.588140 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q8p8\" (UniqueName: \"kubernetes.io/projected/4f19193f-e9fe-4de1-a515-be021e499af8-kube-api-access-8q8p8\") pod \"marketplace-operator-79b997595-rhn6c\" (UID: \"4f19193f-e9fe-4de1-a515-be021e499af8\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhn6c" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.589692 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4f19193f-e9fe-4de1-a515-be021e499af8-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rhn6c\" (UID: \"4f19193f-e9fe-4de1-a515-be021e499af8\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhn6c" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.594341 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4f19193f-e9fe-4de1-a515-be021e499af8-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rhn6c\" (UID: \"4f19193f-e9fe-4de1-a515-be021e499af8\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhn6c" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.602901 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q8p8\" (UniqueName: \"kubernetes.io/projected/4f19193f-e9fe-4de1-a515-be021e499af8-kube-api-access-8q8p8\") pod \"marketplace-operator-79b997595-rhn6c\" (UID: \"4f19193f-e9fe-4de1-a515-be021e499af8\") " pod="openshift-marketplace/marketplace-operator-79b997595-rhn6c" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.724558 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rhn6c" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.826439 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6k6dn" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.857550 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-78cmv" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.880287 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fdrpb" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.882394 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-66hrr" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.890858 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6d137db5-66a8-4066-b8d2-e88adcaaf12c-marketplace-operator-metrics\") pod \"6d137db5-66a8-4066-b8d2-e88adcaaf12c\" (UID: \"6d137db5-66a8-4066-b8d2-e88adcaaf12c\") " Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.890932 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea34a4e8-c842-4a23-933a-8ecd3f882d19-catalog-content\") pod \"ea34a4e8-c842-4a23-933a-8ecd3f882d19\" (UID: \"ea34a4e8-c842-4a23-933a-8ecd3f882d19\") " Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.890972 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea34a4e8-c842-4a23-933a-8ecd3f882d19-utilities\") pod \"ea34a4e8-c842-4a23-933a-8ecd3f882d19\" (UID: \"ea34a4e8-c842-4a23-933a-8ecd3f882d19\") " Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.891008 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l42zk\" (UniqueName: \"kubernetes.io/projected/6d137db5-66a8-4066-b8d2-e88adcaaf12c-kube-api-access-l42zk\") pod \"6d137db5-66a8-4066-b8d2-e88adcaaf12c\" (UID: \"6d137db5-66a8-4066-b8d2-e88adcaaf12c\") " Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.891036 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqmfs\" (UniqueName: \"kubernetes.io/projected/ea34a4e8-c842-4a23-933a-8ecd3f882d19-kube-api-access-vqmfs\") pod \"ea34a4e8-c842-4a23-933a-8ecd3f882d19\" (UID: \"ea34a4e8-c842-4a23-933a-8ecd3f882d19\") " Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.891085 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svl4c\" (UniqueName: \"kubernetes.io/projected/b8150488-b2c9-45e0-8209-9d6a73b64834-kube-api-access-svl4c\") pod \"b8150488-b2c9-45e0-8209-9d6a73b64834\" (UID: \"b8150488-b2c9-45e0-8209-9d6a73b64834\") " Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.891110 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8150488-b2c9-45e0-8209-9d6a73b64834-catalog-content\") pod \"b8150488-b2c9-45e0-8209-9d6a73b64834\" (UID: \"b8150488-b2c9-45e0-8209-9d6a73b64834\") " Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.891134 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8150488-b2c9-45e0-8209-9d6a73b64834-utilities\") pod \"b8150488-b2c9-45e0-8209-9d6a73b64834\" (UID: \"b8150488-b2c9-45e0-8209-9d6a73b64834\") " Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.891156 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fac56d4f-c941-486b-93cb-6931afbd89e0-utilities\") pod \"fac56d4f-c941-486b-93cb-6931afbd89e0\" (UID: \"fac56d4f-c941-486b-93cb-6931afbd89e0\") " Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.891181 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fac56d4f-c941-486b-93cb-6931afbd89e0-catalog-content\") pod \"fac56d4f-c941-486b-93cb-6931afbd89e0\" (UID: \"fac56d4f-c941-486b-93cb-6931afbd89e0\") " Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.891210 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6d137db5-66a8-4066-b8d2-e88adcaaf12c-marketplace-trusted-ca\") pod \"6d137db5-66a8-4066-b8d2-e88adcaaf12c\" (UID: \"6d137db5-66a8-4066-b8d2-e88adcaaf12c\") " Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.891231 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qgh9\" (UniqueName: \"kubernetes.io/projected/fac56d4f-c941-486b-93cb-6931afbd89e0-kube-api-access-2qgh9\") pod \"fac56d4f-c941-486b-93cb-6931afbd89e0\" (UID: \"fac56d4f-c941-486b-93cb-6931afbd89e0\") " Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.892239 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vn2xd" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.892854 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea34a4e8-c842-4a23-933a-8ecd3f882d19-utilities" (OuterVolumeSpecName: "utilities") pod "ea34a4e8-c842-4a23-933a-8ecd3f882d19" (UID: "ea34a4e8-c842-4a23-933a-8ecd3f882d19"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.895283 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fac56d4f-c941-486b-93cb-6931afbd89e0-utilities" (OuterVolumeSpecName: "utilities") pod "fac56d4f-c941-486b-93cb-6931afbd89e0" (UID: "fac56d4f-c941-486b-93cb-6931afbd89e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.897097 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d137db5-66a8-4066-b8d2-e88adcaaf12c-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "6d137db5-66a8-4066-b8d2-e88adcaaf12c" (UID: "6d137db5-66a8-4066-b8d2-e88adcaaf12c"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.898949 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fac56d4f-c941-486b-93cb-6931afbd89e0-kube-api-access-2qgh9" (OuterVolumeSpecName: "kube-api-access-2qgh9") pod "fac56d4f-c941-486b-93cb-6931afbd89e0" (UID: "fac56d4f-c941-486b-93cb-6931afbd89e0"). InnerVolumeSpecName "kube-api-access-2qgh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.900695 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8150488-b2c9-45e0-8209-9d6a73b64834-kube-api-access-svl4c" (OuterVolumeSpecName: "kube-api-access-svl4c") pod "b8150488-b2c9-45e0-8209-9d6a73b64834" (UID: "b8150488-b2c9-45e0-8209-9d6a73b64834"). InnerVolumeSpecName "kube-api-access-svl4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.900853 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d137db5-66a8-4066-b8d2-e88adcaaf12c-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "6d137db5-66a8-4066-b8d2-e88adcaaf12c" (UID: "6d137db5-66a8-4066-b8d2-e88adcaaf12c"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.901071 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d137db5-66a8-4066-b8d2-e88adcaaf12c-kube-api-access-l42zk" (OuterVolumeSpecName: "kube-api-access-l42zk") pod "6d137db5-66a8-4066-b8d2-e88adcaaf12c" (UID: "6d137db5-66a8-4066-b8d2-e88adcaaf12c"). InnerVolumeSpecName "kube-api-access-l42zk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.901703 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8150488-b2c9-45e0-8209-9d6a73b64834-utilities" (OuterVolumeSpecName: "utilities") pod "b8150488-b2c9-45e0-8209-9d6a73b64834" (UID: "b8150488-b2c9-45e0-8209-9d6a73b64834"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.902659 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea34a4e8-c842-4a23-933a-8ecd3f882d19-kube-api-access-vqmfs" (OuterVolumeSpecName: "kube-api-access-vqmfs") pod "ea34a4e8-c842-4a23-933a-8ecd3f882d19" (UID: "ea34a4e8-c842-4a23-933a-8ecd3f882d19"). InnerVolumeSpecName "kube-api-access-vqmfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.924968 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fac56d4f-c941-486b-93cb-6931afbd89e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fac56d4f-c941-486b-93cb-6931afbd89e0" (UID: "fac56d4f-c941-486b-93cb-6931afbd89e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.971083 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8150488-b2c9-45e0-8209-9d6a73b64834-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b8150488-b2c9-45e0-8209-9d6a73b64834" (UID: "b8150488-b2c9-45e0-8209-9d6a73b64834"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.972430 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea34a4e8-c842-4a23-933a-8ecd3f882d19-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ea34a4e8-c842-4a23-933a-8ecd3f882d19" (UID: "ea34a4e8-c842-4a23-933a-8ecd3f882d19"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.992016 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acf903d9-ba78-488a-b55e-6ecd290fe774-catalog-content\") pod \"acf903d9-ba78-488a-b55e-6ecd290fe774\" (UID: \"acf903d9-ba78-488a-b55e-6ecd290fe774\") " Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.992267 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsmhj\" (UniqueName: \"kubernetes.io/projected/acf903d9-ba78-488a-b55e-6ecd290fe774-kube-api-access-dsmhj\") pod \"acf903d9-ba78-488a-b55e-6ecd290fe774\" (UID: \"acf903d9-ba78-488a-b55e-6ecd290fe774\") " Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.992383 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acf903d9-ba78-488a-b55e-6ecd290fe774-utilities\") pod \"acf903d9-ba78-488a-b55e-6ecd290fe774\" (UID: \"acf903d9-ba78-488a-b55e-6ecd290fe774\") " Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.993204 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/acf903d9-ba78-488a-b55e-6ecd290fe774-utilities" (OuterVolumeSpecName: "utilities") pod "acf903d9-ba78-488a-b55e-6ecd290fe774" (UID: "acf903d9-ba78-488a-b55e-6ecd290fe774"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.993320 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea34a4e8-c842-4a23-933a-8ecd3f882d19-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.993342 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea34a4e8-c842-4a23-933a-8ecd3f882d19-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.993352 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l42zk\" (UniqueName: \"kubernetes.io/projected/6d137db5-66a8-4066-b8d2-e88adcaaf12c-kube-api-access-l42zk\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.993363 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqmfs\" (UniqueName: \"kubernetes.io/projected/ea34a4e8-c842-4a23-933a-8ecd3f882d19-kube-api-access-vqmfs\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.993401 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acf903d9-ba78-488a-b55e-6ecd290fe774-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.993414 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svl4c\" (UniqueName: \"kubernetes.io/projected/b8150488-b2c9-45e0-8209-9d6a73b64834-kube-api-access-svl4c\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.993434 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8150488-b2c9-45e0-8209-9d6a73b64834-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.993442 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8150488-b2c9-45e0-8209-9d6a73b64834-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.993452 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fac56d4f-c941-486b-93cb-6931afbd89e0-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.993475 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fac56d4f-c941-486b-93cb-6931afbd89e0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.993513 4934 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6d137db5-66a8-4066-b8d2-e88adcaaf12c-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.993559 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qgh9\" (UniqueName: \"kubernetes.io/projected/fac56d4f-c941-486b-93cb-6931afbd89e0-kube-api-access-2qgh9\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.993624 4934 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6d137db5-66a8-4066-b8d2-e88adcaaf12c-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:51 crc kubenswrapper[4934]: I1002 09:52:51.995646 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acf903d9-ba78-488a-b55e-6ecd290fe774-kube-api-access-dsmhj" (OuterVolumeSpecName: "kube-api-access-dsmhj") pod "acf903d9-ba78-488a-b55e-6ecd290fe774" (UID: "acf903d9-ba78-488a-b55e-6ecd290fe774"). InnerVolumeSpecName "kube-api-access-dsmhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.077225 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/acf903d9-ba78-488a-b55e-6ecd290fe774-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "acf903d9-ba78-488a-b55e-6ecd290fe774" (UID: "acf903d9-ba78-488a-b55e-6ecd290fe774"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.094306 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsmhj\" (UniqueName: \"kubernetes.io/projected/acf903d9-ba78-488a-b55e-6ecd290fe774-kube-api-access-dsmhj\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.094351 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acf903d9-ba78-488a-b55e-6ecd290fe774-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.196343 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rhn6c"] Oct 02 09:52:52 crc kubenswrapper[4934]: W1002 09:52:52.202224 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4f19193f_e9fe_4de1_a515_be021e499af8.slice/crio-f44f511c9b19a3535b4859b1a56e72d1d9807de70f49ec601cb32e5a4ae26ecf WatchSource:0}: Error finding container f44f511c9b19a3535b4859b1a56e72d1d9807de70f49ec601cb32e5a4ae26ecf: Status 404 returned error can't find the container with id f44f511c9b19a3535b4859b1a56e72d1d9807de70f49ec601cb32e5a4ae26ecf Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.508910 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fdrpb" event={"ID":"b8150488-b2c9-45e0-8209-9d6a73b64834","Type":"ContainerDied","Data":"9fd017cdcae70a0f399ea6316cbbfec49f9b1b62aa48164dd181dc43b6172807"} Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.509199 4934 scope.go:117] "RemoveContainer" containerID="00b5632742793c3358edd24ff78e0322fc115b198ac65b62756a1569e9f480f9" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.509419 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fdrpb" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.512640 4934 generic.go:334] "Generic (PLEG): container finished" podID="6d137db5-66a8-4066-b8d2-e88adcaaf12c" containerID="cd49a4c22830bb661a44e96674b3557b609b43ff491066ec8fdf67883616f50d" exitCode=0 Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.512686 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-66hrr" event={"ID":"6d137db5-66a8-4066-b8d2-e88adcaaf12c","Type":"ContainerDied","Data":"cd49a4c22830bb661a44e96674b3557b609b43ff491066ec8fdf67883616f50d"} Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.512719 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-66hrr" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.512742 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-66hrr" event={"ID":"6d137db5-66a8-4066-b8d2-e88adcaaf12c","Type":"ContainerDied","Data":"3e72008b0d74ac2546100da34321d6fd469d741981a96a52127b165bae10326e"} Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.518251 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6k6dn" event={"ID":"ea34a4e8-c842-4a23-933a-8ecd3f882d19","Type":"ContainerDied","Data":"755a3ce9f22b3c84ce99df5512871cb51b4c4c06906ce8a9b9fc7f18c6463e38"} Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.518353 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6k6dn" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.520195 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rhn6c" event={"ID":"4f19193f-e9fe-4de1-a515-be021e499af8","Type":"ContainerStarted","Data":"35a7e1b948fa2a16a96a8b78072d204c89837a1b830ac7146a5f7aa61a8fc739"} Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.520232 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rhn6c" event={"ID":"4f19193f-e9fe-4de1-a515-be021e499af8","Type":"ContainerStarted","Data":"f44f511c9b19a3535b4859b1a56e72d1d9807de70f49ec601cb32e5a4ae26ecf"} Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.521086 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-rhn6c" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.522643 4934 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-rhn6c container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" start-of-body= Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.522686 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-rhn6c" podUID="4f19193f-e9fe-4de1-a515-be021e499af8" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.527980 4934 generic.go:334] "Generic (PLEG): container finished" podID="acf903d9-ba78-488a-b55e-6ecd290fe774" containerID="760adf804fc502d51d5b18c27dad88d647e7fe1ae60a645d8caa715dc86f4341" exitCode=0 Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.528052 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vn2xd" event={"ID":"acf903d9-ba78-488a-b55e-6ecd290fe774","Type":"ContainerDied","Data":"760adf804fc502d51d5b18c27dad88d647e7fe1ae60a645d8caa715dc86f4341"} Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.528079 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vn2xd" event={"ID":"acf903d9-ba78-488a-b55e-6ecd290fe774","Type":"ContainerDied","Data":"1a6331787b0998c2756ee4ca02b34e4d188a6420ff77a517022e7649f87920ae"} Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.528170 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vn2xd" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.540832 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-rhn6c" podStartSLOduration=1.540811231 podStartE2EDuration="1.540811231s" podCreationTimestamp="2025-10-02 09:52:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:52:52.54003685 +0000 UTC m=+244.292678372" watchObservedRunningTime="2025-10-02 09:52:52.540811231 +0000 UTC m=+244.293452753" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.546295 4934 generic.go:334] "Generic (PLEG): container finished" podID="fac56d4f-c941-486b-93cb-6931afbd89e0" containerID="38ea65cbd2658ea8cc3a93bb5838e6fb6c6a3d052166ed430c880aa17f88dba3" exitCode=0 Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.546344 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-78cmv" event={"ID":"fac56d4f-c941-486b-93cb-6931afbd89e0","Type":"ContainerDied","Data":"38ea65cbd2658ea8cc3a93bb5838e6fb6c6a3d052166ed430c880aa17f88dba3"} Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.546368 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-78cmv" event={"ID":"fac56d4f-c941-486b-93cb-6931afbd89e0","Type":"ContainerDied","Data":"0558673d11c10c65ad87f8aa4d3c4277743949fc1e3a8915af0d31fbaca77fa3"} Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.546459 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-78cmv" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.554325 4934 scope.go:117] "RemoveContainer" containerID="f1aea8a08be906a16f7caa992a0e4648b79b5d08dc050ed271c0130f8ebede85" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.582631 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fdrpb"] Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.583290 4934 scope.go:117] "RemoveContainer" containerID="8c7dd1fdadb86aa4fdd32321fb249a4eb56f48f5b7388f0a5c00ea5595ec3dd6" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.587896 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fdrpb"] Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.597404 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-66hrr"] Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.601342 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-66hrr"] Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.601866 4934 scope.go:117] "RemoveContainer" containerID="cd49a4c22830bb661a44e96674b3557b609b43ff491066ec8fdf67883616f50d" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.611558 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6k6dn"] Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.615187 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6k6dn"] Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.620187 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-78cmv"] Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.622883 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-78cmv"] Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.623251 4934 scope.go:117] "RemoveContainer" containerID="cd49a4c22830bb661a44e96674b3557b609b43ff491066ec8fdf67883616f50d" Oct 02 09:52:52 crc kubenswrapper[4934]: E1002 09:52:52.624110 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd49a4c22830bb661a44e96674b3557b609b43ff491066ec8fdf67883616f50d\": container with ID starting with cd49a4c22830bb661a44e96674b3557b609b43ff491066ec8fdf67883616f50d not found: ID does not exist" containerID="cd49a4c22830bb661a44e96674b3557b609b43ff491066ec8fdf67883616f50d" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.624148 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd49a4c22830bb661a44e96674b3557b609b43ff491066ec8fdf67883616f50d"} err="failed to get container status \"cd49a4c22830bb661a44e96674b3557b609b43ff491066ec8fdf67883616f50d\": rpc error: code = NotFound desc = could not find container \"cd49a4c22830bb661a44e96674b3557b609b43ff491066ec8fdf67883616f50d\": container with ID starting with cd49a4c22830bb661a44e96674b3557b609b43ff491066ec8fdf67883616f50d not found: ID does not exist" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.624191 4934 scope.go:117] "RemoveContainer" containerID="5a83f9d75250d2d008490d6b45e41978f650b5e62519f19429d6732dabd03e7c" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.631689 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vn2xd"] Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.639061 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vn2xd"] Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.641272 4934 scope.go:117] "RemoveContainer" containerID="3b963f52f5103d064cce85dfc5d3fedb1aa2c7c534722d0be9ba5c4f5278015e" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.653209 4934 scope.go:117] "RemoveContainer" containerID="d1ed750ce3a57c63d62725b7606b9bc3a083a20ae715646eadbfa54479a8b66b" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.678371 4934 scope.go:117] "RemoveContainer" containerID="760adf804fc502d51d5b18c27dad88d647e7fe1ae60a645d8caa715dc86f4341" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.695230 4934 scope.go:117] "RemoveContainer" containerID="4b4ab48813919caa119eefa5cc6a5f532f7620db21530457bdd66677157e4f77" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.709974 4934 scope.go:117] "RemoveContainer" containerID="e8bb383ed717e61065ad617fbe6a0ee34e1f8a93c8eb038696f510bbdee61dd9" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.723249 4934 scope.go:117] "RemoveContainer" containerID="760adf804fc502d51d5b18c27dad88d647e7fe1ae60a645d8caa715dc86f4341" Oct 02 09:52:52 crc kubenswrapper[4934]: E1002 09:52:52.723731 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"760adf804fc502d51d5b18c27dad88d647e7fe1ae60a645d8caa715dc86f4341\": container with ID starting with 760adf804fc502d51d5b18c27dad88d647e7fe1ae60a645d8caa715dc86f4341 not found: ID does not exist" containerID="760adf804fc502d51d5b18c27dad88d647e7fe1ae60a645d8caa715dc86f4341" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.723768 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"760adf804fc502d51d5b18c27dad88d647e7fe1ae60a645d8caa715dc86f4341"} err="failed to get container status \"760adf804fc502d51d5b18c27dad88d647e7fe1ae60a645d8caa715dc86f4341\": rpc error: code = NotFound desc = could not find container \"760adf804fc502d51d5b18c27dad88d647e7fe1ae60a645d8caa715dc86f4341\": container with ID starting with 760adf804fc502d51d5b18c27dad88d647e7fe1ae60a645d8caa715dc86f4341 not found: ID does not exist" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.723804 4934 scope.go:117] "RemoveContainer" containerID="4b4ab48813919caa119eefa5cc6a5f532f7620db21530457bdd66677157e4f77" Oct 02 09:52:52 crc kubenswrapper[4934]: E1002 09:52:52.724096 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b4ab48813919caa119eefa5cc6a5f532f7620db21530457bdd66677157e4f77\": container with ID starting with 4b4ab48813919caa119eefa5cc6a5f532f7620db21530457bdd66677157e4f77 not found: ID does not exist" containerID="4b4ab48813919caa119eefa5cc6a5f532f7620db21530457bdd66677157e4f77" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.724120 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b4ab48813919caa119eefa5cc6a5f532f7620db21530457bdd66677157e4f77"} err="failed to get container status \"4b4ab48813919caa119eefa5cc6a5f532f7620db21530457bdd66677157e4f77\": rpc error: code = NotFound desc = could not find container \"4b4ab48813919caa119eefa5cc6a5f532f7620db21530457bdd66677157e4f77\": container with ID starting with 4b4ab48813919caa119eefa5cc6a5f532f7620db21530457bdd66677157e4f77 not found: ID does not exist" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.724134 4934 scope.go:117] "RemoveContainer" containerID="e8bb383ed717e61065ad617fbe6a0ee34e1f8a93c8eb038696f510bbdee61dd9" Oct 02 09:52:52 crc kubenswrapper[4934]: E1002 09:52:52.724361 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8bb383ed717e61065ad617fbe6a0ee34e1f8a93c8eb038696f510bbdee61dd9\": container with ID starting with e8bb383ed717e61065ad617fbe6a0ee34e1f8a93c8eb038696f510bbdee61dd9 not found: ID does not exist" containerID="e8bb383ed717e61065ad617fbe6a0ee34e1f8a93c8eb038696f510bbdee61dd9" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.724378 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8bb383ed717e61065ad617fbe6a0ee34e1f8a93c8eb038696f510bbdee61dd9"} err="failed to get container status \"e8bb383ed717e61065ad617fbe6a0ee34e1f8a93c8eb038696f510bbdee61dd9\": rpc error: code = NotFound desc = could not find container \"e8bb383ed717e61065ad617fbe6a0ee34e1f8a93c8eb038696f510bbdee61dd9\": container with ID starting with e8bb383ed717e61065ad617fbe6a0ee34e1f8a93c8eb038696f510bbdee61dd9 not found: ID does not exist" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.724389 4934 scope.go:117] "RemoveContainer" containerID="38ea65cbd2658ea8cc3a93bb5838e6fb6c6a3d052166ed430c880aa17f88dba3" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.736159 4934 scope.go:117] "RemoveContainer" containerID="1ab4526e8d685fd70ffe68ee552c9061d0b65459323e4aea17a8020258448b84" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.747714 4934 scope.go:117] "RemoveContainer" containerID="7b7e05d4289a96d7ab8c886aae97669defac87fa69c970a55928de92875c792d" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.762747 4934 scope.go:117] "RemoveContainer" containerID="38ea65cbd2658ea8cc3a93bb5838e6fb6c6a3d052166ed430c880aa17f88dba3" Oct 02 09:52:52 crc kubenswrapper[4934]: E1002 09:52:52.763097 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38ea65cbd2658ea8cc3a93bb5838e6fb6c6a3d052166ed430c880aa17f88dba3\": container with ID starting with 38ea65cbd2658ea8cc3a93bb5838e6fb6c6a3d052166ed430c880aa17f88dba3 not found: ID does not exist" containerID="38ea65cbd2658ea8cc3a93bb5838e6fb6c6a3d052166ed430c880aa17f88dba3" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.763129 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38ea65cbd2658ea8cc3a93bb5838e6fb6c6a3d052166ed430c880aa17f88dba3"} err="failed to get container status \"38ea65cbd2658ea8cc3a93bb5838e6fb6c6a3d052166ed430c880aa17f88dba3\": rpc error: code = NotFound desc = could not find container \"38ea65cbd2658ea8cc3a93bb5838e6fb6c6a3d052166ed430c880aa17f88dba3\": container with ID starting with 38ea65cbd2658ea8cc3a93bb5838e6fb6c6a3d052166ed430c880aa17f88dba3 not found: ID does not exist" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.763156 4934 scope.go:117] "RemoveContainer" containerID="1ab4526e8d685fd70ffe68ee552c9061d0b65459323e4aea17a8020258448b84" Oct 02 09:52:52 crc kubenswrapper[4934]: E1002 09:52:52.763507 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ab4526e8d685fd70ffe68ee552c9061d0b65459323e4aea17a8020258448b84\": container with ID starting with 1ab4526e8d685fd70ffe68ee552c9061d0b65459323e4aea17a8020258448b84 not found: ID does not exist" containerID="1ab4526e8d685fd70ffe68ee552c9061d0b65459323e4aea17a8020258448b84" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.763531 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ab4526e8d685fd70ffe68ee552c9061d0b65459323e4aea17a8020258448b84"} err="failed to get container status \"1ab4526e8d685fd70ffe68ee552c9061d0b65459323e4aea17a8020258448b84\": rpc error: code = NotFound desc = could not find container \"1ab4526e8d685fd70ffe68ee552c9061d0b65459323e4aea17a8020258448b84\": container with ID starting with 1ab4526e8d685fd70ffe68ee552c9061d0b65459323e4aea17a8020258448b84 not found: ID does not exist" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.763548 4934 scope.go:117] "RemoveContainer" containerID="7b7e05d4289a96d7ab8c886aae97669defac87fa69c970a55928de92875c792d" Oct 02 09:52:52 crc kubenswrapper[4934]: E1002 09:52:52.763790 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b7e05d4289a96d7ab8c886aae97669defac87fa69c970a55928de92875c792d\": container with ID starting with 7b7e05d4289a96d7ab8c886aae97669defac87fa69c970a55928de92875c792d not found: ID does not exist" containerID="7b7e05d4289a96d7ab8c886aae97669defac87fa69c970a55928de92875c792d" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.763813 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b7e05d4289a96d7ab8c886aae97669defac87fa69c970a55928de92875c792d"} err="failed to get container status \"7b7e05d4289a96d7ab8c886aae97669defac87fa69c970a55928de92875c792d\": rpc error: code = NotFound desc = could not find container \"7b7e05d4289a96d7ab8c886aae97669defac87fa69c970a55928de92875c792d\": container with ID starting with 7b7e05d4289a96d7ab8c886aae97669defac87fa69c970a55928de92875c792d not found: ID does not exist" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.919883 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d137db5-66a8-4066-b8d2-e88adcaaf12c" path="/var/lib/kubelet/pods/6d137db5-66a8-4066-b8d2-e88adcaaf12c/volumes" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.920438 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acf903d9-ba78-488a-b55e-6ecd290fe774" path="/var/lib/kubelet/pods/acf903d9-ba78-488a-b55e-6ecd290fe774/volumes" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.921136 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8150488-b2c9-45e0-8209-9d6a73b64834" path="/var/lib/kubelet/pods/b8150488-b2c9-45e0-8209-9d6a73b64834/volumes" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.922187 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea34a4e8-c842-4a23-933a-8ecd3f882d19" path="/var/lib/kubelet/pods/ea34a4e8-c842-4a23-933a-8ecd3f882d19/volumes" Oct 02 09:52:52 crc kubenswrapper[4934]: I1002 09:52:52.922756 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fac56d4f-c941-486b-93cb-6931afbd89e0" path="/var/lib/kubelet/pods/fac56d4f-c941-486b-93cb-6931afbd89e0/volumes" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.557570 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-rhn6c" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.581808 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4bvl5"] Oct 02 09:52:53 crc kubenswrapper[4934]: E1002 09:52:53.582004 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea34a4e8-c842-4a23-933a-8ecd3f882d19" containerName="extract-content" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.582015 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea34a4e8-c842-4a23-933a-8ecd3f882d19" containerName="extract-content" Oct 02 09:52:53 crc kubenswrapper[4934]: E1002 09:52:53.582023 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8150488-b2c9-45e0-8209-9d6a73b64834" containerName="extract-utilities" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.582031 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8150488-b2c9-45e0-8209-9d6a73b64834" containerName="extract-utilities" Oct 02 09:52:53 crc kubenswrapper[4934]: E1002 09:52:53.582038 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acf903d9-ba78-488a-b55e-6ecd290fe774" containerName="registry-server" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.582043 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="acf903d9-ba78-488a-b55e-6ecd290fe774" containerName="registry-server" Oct 02 09:52:53 crc kubenswrapper[4934]: E1002 09:52:53.582054 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8150488-b2c9-45e0-8209-9d6a73b64834" containerName="extract-content" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.582061 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8150488-b2c9-45e0-8209-9d6a73b64834" containerName="extract-content" Oct 02 09:52:53 crc kubenswrapper[4934]: E1002 09:52:53.582071 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea34a4e8-c842-4a23-933a-8ecd3f882d19" containerName="registry-server" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.582076 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea34a4e8-c842-4a23-933a-8ecd3f882d19" containerName="registry-server" Oct 02 09:52:53 crc kubenswrapper[4934]: E1002 09:52:53.582084 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d137db5-66a8-4066-b8d2-e88adcaaf12c" containerName="marketplace-operator" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.582090 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d137db5-66a8-4066-b8d2-e88adcaaf12c" containerName="marketplace-operator" Oct 02 09:52:53 crc kubenswrapper[4934]: E1002 09:52:53.582096 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8150488-b2c9-45e0-8209-9d6a73b64834" containerName="registry-server" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.582104 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8150488-b2c9-45e0-8209-9d6a73b64834" containerName="registry-server" Oct 02 09:52:53 crc kubenswrapper[4934]: E1002 09:52:53.582113 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fac56d4f-c941-486b-93cb-6931afbd89e0" containerName="registry-server" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.582118 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="fac56d4f-c941-486b-93cb-6931afbd89e0" containerName="registry-server" Oct 02 09:52:53 crc kubenswrapper[4934]: E1002 09:52:53.582126 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acf903d9-ba78-488a-b55e-6ecd290fe774" containerName="extract-utilities" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.582132 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="acf903d9-ba78-488a-b55e-6ecd290fe774" containerName="extract-utilities" Oct 02 09:52:53 crc kubenswrapper[4934]: E1002 09:52:53.582140 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea34a4e8-c842-4a23-933a-8ecd3f882d19" containerName="extract-utilities" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.582145 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea34a4e8-c842-4a23-933a-8ecd3f882d19" containerName="extract-utilities" Oct 02 09:52:53 crc kubenswrapper[4934]: E1002 09:52:53.582153 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fac56d4f-c941-486b-93cb-6931afbd89e0" containerName="extract-utilities" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.582159 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="fac56d4f-c941-486b-93cb-6931afbd89e0" containerName="extract-utilities" Oct 02 09:52:53 crc kubenswrapper[4934]: E1002 09:52:53.582166 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fac56d4f-c941-486b-93cb-6931afbd89e0" containerName="extract-content" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.582171 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="fac56d4f-c941-486b-93cb-6931afbd89e0" containerName="extract-content" Oct 02 09:52:53 crc kubenswrapper[4934]: E1002 09:52:53.582178 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acf903d9-ba78-488a-b55e-6ecd290fe774" containerName="extract-content" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.582183 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="acf903d9-ba78-488a-b55e-6ecd290fe774" containerName="extract-content" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.582255 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="acf903d9-ba78-488a-b55e-6ecd290fe774" containerName="registry-server" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.582265 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="fac56d4f-c941-486b-93cb-6931afbd89e0" containerName="registry-server" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.582272 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8150488-b2c9-45e0-8209-9d6a73b64834" containerName="registry-server" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.582282 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d137db5-66a8-4066-b8d2-e88adcaaf12c" containerName="marketplace-operator" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.582292 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea34a4e8-c842-4a23-933a-8ecd3f882d19" containerName="registry-server" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.582948 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4bvl5" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.584570 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.606336 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4bvl5"] Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.608357 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hswqb\" (UniqueName: \"kubernetes.io/projected/b89bbfbf-3398-44a8-a54b-3fc5b55bec21-kube-api-access-hswqb\") pod \"redhat-marketplace-4bvl5\" (UID: \"b89bbfbf-3398-44a8-a54b-3fc5b55bec21\") " pod="openshift-marketplace/redhat-marketplace-4bvl5" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.608555 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b89bbfbf-3398-44a8-a54b-3fc5b55bec21-catalog-content\") pod \"redhat-marketplace-4bvl5\" (UID: \"b89bbfbf-3398-44a8-a54b-3fc5b55bec21\") " pod="openshift-marketplace/redhat-marketplace-4bvl5" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.608655 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b89bbfbf-3398-44a8-a54b-3fc5b55bec21-utilities\") pod \"redhat-marketplace-4bvl5\" (UID: \"b89bbfbf-3398-44a8-a54b-3fc5b55bec21\") " pod="openshift-marketplace/redhat-marketplace-4bvl5" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.710283 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b89bbfbf-3398-44a8-a54b-3fc5b55bec21-utilities\") pod \"redhat-marketplace-4bvl5\" (UID: \"b89bbfbf-3398-44a8-a54b-3fc5b55bec21\") " pod="openshift-marketplace/redhat-marketplace-4bvl5" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.710436 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hswqb\" (UniqueName: \"kubernetes.io/projected/b89bbfbf-3398-44a8-a54b-3fc5b55bec21-kube-api-access-hswqb\") pod \"redhat-marketplace-4bvl5\" (UID: \"b89bbfbf-3398-44a8-a54b-3fc5b55bec21\") " pod="openshift-marketplace/redhat-marketplace-4bvl5" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.710487 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b89bbfbf-3398-44a8-a54b-3fc5b55bec21-catalog-content\") pod \"redhat-marketplace-4bvl5\" (UID: \"b89bbfbf-3398-44a8-a54b-3fc5b55bec21\") " pod="openshift-marketplace/redhat-marketplace-4bvl5" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.710809 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b89bbfbf-3398-44a8-a54b-3fc5b55bec21-utilities\") pod \"redhat-marketplace-4bvl5\" (UID: \"b89bbfbf-3398-44a8-a54b-3fc5b55bec21\") " pod="openshift-marketplace/redhat-marketplace-4bvl5" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.710915 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b89bbfbf-3398-44a8-a54b-3fc5b55bec21-catalog-content\") pod \"redhat-marketplace-4bvl5\" (UID: \"b89bbfbf-3398-44a8-a54b-3fc5b55bec21\") " pod="openshift-marketplace/redhat-marketplace-4bvl5" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.729336 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hswqb\" (UniqueName: \"kubernetes.io/projected/b89bbfbf-3398-44a8-a54b-3fc5b55bec21-kube-api-access-hswqb\") pod \"redhat-marketplace-4bvl5\" (UID: \"b89bbfbf-3398-44a8-a54b-3fc5b55bec21\") " pod="openshift-marketplace/redhat-marketplace-4bvl5" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.785734 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vk4jw"] Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.786904 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vk4jw" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.789794 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.795477 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vk4jw"] Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.904867 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4bvl5" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.912492 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/599deecc-b859-4fc9-bb70-5675ef509325-utilities\") pod \"redhat-operators-vk4jw\" (UID: \"599deecc-b859-4fc9-bb70-5675ef509325\") " pod="openshift-marketplace/redhat-operators-vk4jw" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.912596 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/599deecc-b859-4fc9-bb70-5675ef509325-catalog-content\") pod \"redhat-operators-vk4jw\" (UID: \"599deecc-b859-4fc9-bb70-5675ef509325\") " pod="openshift-marketplace/redhat-operators-vk4jw" Oct 02 09:52:53 crc kubenswrapper[4934]: I1002 09:52:53.912658 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lck9d\" (UniqueName: \"kubernetes.io/projected/599deecc-b859-4fc9-bb70-5675ef509325-kube-api-access-lck9d\") pod \"redhat-operators-vk4jw\" (UID: \"599deecc-b859-4fc9-bb70-5675ef509325\") " pod="openshift-marketplace/redhat-operators-vk4jw" Oct 02 09:52:54 crc kubenswrapper[4934]: I1002 09:52:54.013168 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lck9d\" (UniqueName: \"kubernetes.io/projected/599deecc-b859-4fc9-bb70-5675ef509325-kube-api-access-lck9d\") pod \"redhat-operators-vk4jw\" (UID: \"599deecc-b859-4fc9-bb70-5675ef509325\") " pod="openshift-marketplace/redhat-operators-vk4jw" Oct 02 09:52:54 crc kubenswrapper[4934]: I1002 09:52:54.013465 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/599deecc-b859-4fc9-bb70-5675ef509325-utilities\") pod \"redhat-operators-vk4jw\" (UID: \"599deecc-b859-4fc9-bb70-5675ef509325\") " pod="openshift-marketplace/redhat-operators-vk4jw" Oct 02 09:52:54 crc kubenswrapper[4934]: I1002 09:52:54.013518 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/599deecc-b859-4fc9-bb70-5675ef509325-catalog-content\") pod \"redhat-operators-vk4jw\" (UID: \"599deecc-b859-4fc9-bb70-5675ef509325\") " pod="openshift-marketplace/redhat-operators-vk4jw" Oct 02 09:52:54 crc kubenswrapper[4934]: I1002 09:52:54.014174 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/599deecc-b859-4fc9-bb70-5675ef509325-catalog-content\") pod \"redhat-operators-vk4jw\" (UID: \"599deecc-b859-4fc9-bb70-5675ef509325\") " pod="openshift-marketplace/redhat-operators-vk4jw" Oct 02 09:52:54 crc kubenswrapper[4934]: I1002 09:52:54.014260 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/599deecc-b859-4fc9-bb70-5675ef509325-utilities\") pod \"redhat-operators-vk4jw\" (UID: \"599deecc-b859-4fc9-bb70-5675ef509325\") " pod="openshift-marketplace/redhat-operators-vk4jw" Oct 02 09:52:54 crc kubenswrapper[4934]: I1002 09:52:54.031567 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lck9d\" (UniqueName: \"kubernetes.io/projected/599deecc-b859-4fc9-bb70-5675ef509325-kube-api-access-lck9d\") pod \"redhat-operators-vk4jw\" (UID: \"599deecc-b859-4fc9-bb70-5675ef509325\") " pod="openshift-marketplace/redhat-operators-vk4jw" Oct 02 09:52:54 crc kubenswrapper[4934]: I1002 09:52:54.104456 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vk4jw" Oct 02 09:52:54 crc kubenswrapper[4934]: I1002 09:52:54.311229 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4bvl5"] Oct 02 09:52:54 crc kubenswrapper[4934]: I1002 09:52:54.486692 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vk4jw"] Oct 02 09:52:54 crc kubenswrapper[4934]: I1002 09:52:54.561005 4934 generic.go:334] "Generic (PLEG): container finished" podID="b89bbfbf-3398-44a8-a54b-3fc5b55bec21" containerID="46f4d86b47697b19fe835318c09716aa7a7058fa3848026fbb1cca379a56abc5" exitCode=0 Oct 02 09:52:54 crc kubenswrapper[4934]: I1002 09:52:54.561060 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4bvl5" event={"ID":"b89bbfbf-3398-44a8-a54b-3fc5b55bec21","Type":"ContainerDied","Data":"46f4d86b47697b19fe835318c09716aa7a7058fa3848026fbb1cca379a56abc5"} Oct 02 09:52:54 crc kubenswrapper[4934]: I1002 09:52:54.561087 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4bvl5" event={"ID":"b89bbfbf-3398-44a8-a54b-3fc5b55bec21","Type":"ContainerStarted","Data":"7d2af2a75367faaad8e0185289dff33bde601b168d95575238b9d7f5f97b9c57"} Oct 02 09:52:54 crc kubenswrapper[4934]: I1002 09:52:54.563491 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vk4jw" event={"ID":"599deecc-b859-4fc9-bb70-5675ef509325","Type":"ContainerStarted","Data":"cc05328fd55b89c0d1b145fb714e6ae44751739861940dd3403f5fb1526c8d3a"} Oct 02 09:52:55 crc kubenswrapper[4934]: I1002 09:52:55.576771 4934 generic.go:334] "Generic (PLEG): container finished" podID="b89bbfbf-3398-44a8-a54b-3fc5b55bec21" containerID="01a642a11f4d2f7e9a90a9ae6642dcf1c2b6335f1624f5401730ebd9567decbe" exitCode=0 Oct 02 09:52:55 crc kubenswrapper[4934]: I1002 09:52:55.576883 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4bvl5" event={"ID":"b89bbfbf-3398-44a8-a54b-3fc5b55bec21","Type":"ContainerDied","Data":"01a642a11f4d2f7e9a90a9ae6642dcf1c2b6335f1624f5401730ebd9567decbe"} Oct 02 09:52:55 crc kubenswrapper[4934]: I1002 09:52:55.579026 4934 generic.go:334] "Generic (PLEG): container finished" podID="599deecc-b859-4fc9-bb70-5675ef509325" containerID="e88f0ae012628462d577e114b24936d848b8e257dccff79d18cfcadc672e68f4" exitCode=0 Oct 02 09:52:55 crc kubenswrapper[4934]: I1002 09:52:55.580116 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vk4jw" event={"ID":"599deecc-b859-4fc9-bb70-5675ef509325","Type":"ContainerDied","Data":"e88f0ae012628462d577e114b24936d848b8e257dccff79d18cfcadc672e68f4"} Oct 02 09:52:55 crc kubenswrapper[4934]: I1002 09:52:55.980319 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zxq5v"] Oct 02 09:52:55 crc kubenswrapper[4934]: I1002 09:52:55.981239 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zxq5v" Oct 02 09:52:55 crc kubenswrapper[4934]: I1002 09:52:55.985120 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.002063 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zxq5v"] Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.145534 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31f0c5ea-f990-45d1-b501-53ce618fac45-utilities\") pod \"community-operators-zxq5v\" (UID: \"31f0c5ea-f990-45d1-b501-53ce618fac45\") " pod="openshift-marketplace/community-operators-zxq5v" Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.145833 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slk47\" (UniqueName: \"kubernetes.io/projected/31f0c5ea-f990-45d1-b501-53ce618fac45-kube-api-access-slk47\") pod \"community-operators-zxq5v\" (UID: \"31f0c5ea-f990-45d1-b501-53ce618fac45\") " pod="openshift-marketplace/community-operators-zxq5v" Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.146123 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31f0c5ea-f990-45d1-b501-53ce618fac45-catalog-content\") pod \"community-operators-zxq5v\" (UID: \"31f0c5ea-f990-45d1-b501-53ce618fac45\") " pod="openshift-marketplace/community-operators-zxq5v" Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.182069 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sfhmd"] Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.183251 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sfhmd" Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.184722 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.193843 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sfhmd"] Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.247547 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31f0c5ea-f990-45d1-b501-53ce618fac45-catalog-content\") pod \"community-operators-zxq5v\" (UID: \"31f0c5ea-f990-45d1-b501-53ce618fac45\") " pod="openshift-marketplace/community-operators-zxq5v" Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.247641 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31f0c5ea-f990-45d1-b501-53ce618fac45-utilities\") pod \"community-operators-zxq5v\" (UID: \"31f0c5ea-f990-45d1-b501-53ce618fac45\") " pod="openshift-marketplace/community-operators-zxq5v" Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.247668 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slk47\" (UniqueName: \"kubernetes.io/projected/31f0c5ea-f990-45d1-b501-53ce618fac45-kube-api-access-slk47\") pod \"community-operators-zxq5v\" (UID: \"31f0c5ea-f990-45d1-b501-53ce618fac45\") " pod="openshift-marketplace/community-operators-zxq5v" Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.248391 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31f0c5ea-f990-45d1-b501-53ce618fac45-catalog-content\") pod \"community-operators-zxq5v\" (UID: \"31f0c5ea-f990-45d1-b501-53ce618fac45\") " pod="openshift-marketplace/community-operators-zxq5v" Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.248463 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31f0c5ea-f990-45d1-b501-53ce618fac45-utilities\") pod \"community-operators-zxq5v\" (UID: \"31f0c5ea-f990-45d1-b501-53ce618fac45\") " pod="openshift-marketplace/community-operators-zxq5v" Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.267506 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slk47\" (UniqueName: \"kubernetes.io/projected/31f0c5ea-f990-45d1-b501-53ce618fac45-kube-api-access-slk47\") pod \"community-operators-zxq5v\" (UID: \"31f0c5ea-f990-45d1-b501-53ce618fac45\") " pod="openshift-marketplace/community-operators-zxq5v" Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.307161 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zxq5v" Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.349071 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1df910b6-adca-48c6-8508-0e6535c84f6a-utilities\") pod \"certified-operators-sfhmd\" (UID: \"1df910b6-adca-48c6-8508-0e6535c84f6a\") " pod="openshift-marketplace/certified-operators-sfhmd" Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.349152 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84dp5\" (UniqueName: \"kubernetes.io/projected/1df910b6-adca-48c6-8508-0e6535c84f6a-kube-api-access-84dp5\") pod \"certified-operators-sfhmd\" (UID: \"1df910b6-adca-48c6-8508-0e6535c84f6a\") " pod="openshift-marketplace/certified-operators-sfhmd" Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.349522 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1df910b6-adca-48c6-8508-0e6535c84f6a-catalog-content\") pod \"certified-operators-sfhmd\" (UID: \"1df910b6-adca-48c6-8508-0e6535c84f6a\") " pod="openshift-marketplace/certified-operators-sfhmd" Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.451295 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1df910b6-adca-48c6-8508-0e6535c84f6a-catalog-content\") pod \"certified-operators-sfhmd\" (UID: \"1df910b6-adca-48c6-8508-0e6535c84f6a\") " pod="openshift-marketplace/certified-operators-sfhmd" Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.451775 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1df910b6-adca-48c6-8508-0e6535c84f6a-utilities\") pod \"certified-operators-sfhmd\" (UID: \"1df910b6-adca-48c6-8508-0e6535c84f6a\") " pod="openshift-marketplace/certified-operators-sfhmd" Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.451806 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84dp5\" (UniqueName: \"kubernetes.io/projected/1df910b6-adca-48c6-8508-0e6535c84f6a-kube-api-access-84dp5\") pod \"certified-operators-sfhmd\" (UID: \"1df910b6-adca-48c6-8508-0e6535c84f6a\") " pod="openshift-marketplace/certified-operators-sfhmd" Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.451973 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1df910b6-adca-48c6-8508-0e6535c84f6a-catalog-content\") pod \"certified-operators-sfhmd\" (UID: \"1df910b6-adca-48c6-8508-0e6535c84f6a\") " pod="openshift-marketplace/certified-operators-sfhmd" Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.452244 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1df910b6-adca-48c6-8508-0e6535c84f6a-utilities\") pod \"certified-operators-sfhmd\" (UID: \"1df910b6-adca-48c6-8508-0e6535c84f6a\") " pod="openshift-marketplace/certified-operators-sfhmd" Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.474839 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84dp5\" (UniqueName: \"kubernetes.io/projected/1df910b6-adca-48c6-8508-0e6535c84f6a-kube-api-access-84dp5\") pod \"certified-operators-sfhmd\" (UID: \"1df910b6-adca-48c6-8508-0e6535c84f6a\") " pod="openshift-marketplace/certified-operators-sfhmd" Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.502105 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sfhmd" Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.590530 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vk4jw" event={"ID":"599deecc-b859-4fc9-bb70-5675ef509325","Type":"ContainerStarted","Data":"e56d4f1b111c0cedcc30deb109ce7de56354cb02943f16f28d7a984dd0952a88"} Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.607852 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4bvl5" event={"ID":"b89bbfbf-3398-44a8-a54b-3fc5b55bec21","Type":"ContainerStarted","Data":"4d9c3efa4805207e7d768ac7eb56261212c69aaf655faadb3df41e407d918217"} Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.637046 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4bvl5" podStartSLOduration=2.170539772 podStartE2EDuration="3.637023224s" podCreationTimestamp="2025-10-02 09:52:53 +0000 UTC" firstStartedPulling="2025-10-02 09:52:54.56280453 +0000 UTC m=+246.315446052" lastFinishedPulling="2025-10-02 09:52:56.029287982 +0000 UTC m=+247.781929504" observedRunningTime="2025-10-02 09:52:56.634337563 +0000 UTC m=+248.386979085" watchObservedRunningTime="2025-10-02 09:52:56.637023224 +0000 UTC m=+248.389664746" Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.714621 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zxq5v"] Oct 02 09:52:56 crc kubenswrapper[4934]: W1002 09:52:56.722438 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31f0c5ea_f990_45d1_b501_53ce618fac45.slice/crio-04819bf48c31aa109ac2149d491b0dd9962457c1023691601680a3607a53f221 WatchSource:0}: Error finding container 04819bf48c31aa109ac2149d491b0dd9962457c1023691601680a3607a53f221: Status 404 returned error can't find the container with id 04819bf48c31aa109ac2149d491b0dd9962457c1023691601680a3607a53f221 Oct 02 09:52:56 crc kubenswrapper[4934]: I1002 09:52:56.906558 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sfhmd"] Oct 02 09:52:56 crc kubenswrapper[4934]: W1002 09:52:56.998719 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1df910b6_adca_48c6_8508_0e6535c84f6a.slice/crio-8c7dd731e70dedb855ad55448406de4682ca2606628f74c26527c47a92763bde WatchSource:0}: Error finding container 8c7dd731e70dedb855ad55448406de4682ca2606628f74c26527c47a92763bde: Status 404 returned error can't find the container with id 8c7dd731e70dedb855ad55448406de4682ca2606628f74c26527c47a92763bde Oct 02 09:52:57 crc kubenswrapper[4934]: I1002 09:52:57.616847 4934 generic.go:334] "Generic (PLEG): container finished" podID="31f0c5ea-f990-45d1-b501-53ce618fac45" containerID="009257a50c99cd56d866e19fad379787895f19af7add5bf3f52f0ff9c80145f7" exitCode=0 Oct 02 09:52:57 crc kubenswrapper[4934]: I1002 09:52:57.616930 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zxq5v" event={"ID":"31f0c5ea-f990-45d1-b501-53ce618fac45","Type":"ContainerDied","Data":"009257a50c99cd56d866e19fad379787895f19af7add5bf3f52f0ff9c80145f7"} Oct 02 09:52:57 crc kubenswrapper[4934]: I1002 09:52:57.617268 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zxq5v" event={"ID":"31f0c5ea-f990-45d1-b501-53ce618fac45","Type":"ContainerStarted","Data":"04819bf48c31aa109ac2149d491b0dd9962457c1023691601680a3607a53f221"} Oct 02 09:52:57 crc kubenswrapper[4934]: I1002 09:52:57.619512 4934 generic.go:334] "Generic (PLEG): container finished" podID="599deecc-b859-4fc9-bb70-5675ef509325" containerID="e56d4f1b111c0cedcc30deb109ce7de56354cb02943f16f28d7a984dd0952a88" exitCode=0 Oct 02 09:52:57 crc kubenswrapper[4934]: I1002 09:52:57.619568 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vk4jw" event={"ID":"599deecc-b859-4fc9-bb70-5675ef509325","Type":"ContainerDied","Data":"e56d4f1b111c0cedcc30deb109ce7de56354cb02943f16f28d7a984dd0952a88"} Oct 02 09:52:57 crc kubenswrapper[4934]: I1002 09:52:57.625461 4934 generic.go:334] "Generic (PLEG): container finished" podID="1df910b6-adca-48c6-8508-0e6535c84f6a" containerID="7ddedb743564237b763f8a811de1fe9bb378067e9e0046d754f9cddb13ba77ee" exitCode=0 Oct 02 09:52:57 crc kubenswrapper[4934]: I1002 09:52:57.625663 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sfhmd" event={"ID":"1df910b6-adca-48c6-8508-0e6535c84f6a","Type":"ContainerDied","Data":"7ddedb743564237b763f8a811de1fe9bb378067e9e0046d754f9cddb13ba77ee"} Oct 02 09:52:57 crc kubenswrapper[4934]: I1002 09:52:57.625777 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sfhmd" event={"ID":"1df910b6-adca-48c6-8508-0e6535c84f6a","Type":"ContainerStarted","Data":"8c7dd731e70dedb855ad55448406de4682ca2606628f74c26527c47a92763bde"} Oct 02 09:52:58 crc kubenswrapper[4934]: I1002 09:52:58.632899 4934 generic.go:334] "Generic (PLEG): container finished" podID="31f0c5ea-f990-45d1-b501-53ce618fac45" containerID="d8db3ef1616ac0a0f43b51fec0cc858cd6bcdf2a98034743d27660e8eaa92bfe" exitCode=0 Oct 02 09:52:58 crc kubenswrapper[4934]: I1002 09:52:58.633047 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zxq5v" event={"ID":"31f0c5ea-f990-45d1-b501-53ce618fac45","Type":"ContainerDied","Data":"d8db3ef1616ac0a0f43b51fec0cc858cd6bcdf2a98034743d27660e8eaa92bfe"} Oct 02 09:52:58 crc kubenswrapper[4934]: I1002 09:52:58.638097 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vk4jw" event={"ID":"599deecc-b859-4fc9-bb70-5675ef509325","Type":"ContainerStarted","Data":"9e270269202dacda0149bcdcdf11c3685129970f11eb3217a562c495e9a2982f"} Oct 02 09:52:58 crc kubenswrapper[4934]: I1002 09:52:58.639762 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sfhmd" event={"ID":"1df910b6-adca-48c6-8508-0e6535c84f6a","Type":"ContainerStarted","Data":"6b7f02b706290fd1fe8c2d346cdd9ae638d366bb748c4819c6bf6170ab50ab63"} Oct 02 09:52:58 crc kubenswrapper[4934]: I1002 09:52:58.685704 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vk4jw" podStartSLOduration=3.231474355 podStartE2EDuration="5.685679304s" podCreationTimestamp="2025-10-02 09:52:53 +0000 UTC" firstStartedPulling="2025-10-02 09:52:55.581218376 +0000 UTC m=+247.333859908" lastFinishedPulling="2025-10-02 09:52:58.035423335 +0000 UTC m=+249.788064857" observedRunningTime="2025-10-02 09:52:58.663523812 +0000 UTC m=+250.416165334" watchObservedRunningTime="2025-10-02 09:52:58.685679304 +0000 UTC m=+250.438320836" Oct 02 09:52:59 crc kubenswrapper[4934]: I1002 09:52:59.647338 4934 generic.go:334] "Generic (PLEG): container finished" podID="1df910b6-adca-48c6-8508-0e6535c84f6a" containerID="6b7f02b706290fd1fe8c2d346cdd9ae638d366bb748c4819c6bf6170ab50ab63" exitCode=0 Oct 02 09:52:59 crc kubenswrapper[4934]: I1002 09:52:59.647414 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sfhmd" event={"ID":"1df910b6-adca-48c6-8508-0e6535c84f6a","Type":"ContainerDied","Data":"6b7f02b706290fd1fe8c2d346cdd9ae638d366bb748c4819c6bf6170ab50ab63"} Oct 02 09:53:00 crc kubenswrapper[4934]: I1002 09:53:00.655567 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zxq5v" event={"ID":"31f0c5ea-f990-45d1-b501-53ce618fac45","Type":"ContainerStarted","Data":"faadb42164652130f2b90fdc064785b7e9d81060188582fb5ddc8d3c9e9e1672"} Oct 02 09:53:00 crc kubenswrapper[4934]: I1002 09:53:00.657262 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sfhmd" event={"ID":"1df910b6-adca-48c6-8508-0e6535c84f6a","Type":"ContainerStarted","Data":"ae0c202f3ad950753b552300ee7c7f90e43f0d0e2f1854f70c82fd897d1d9c63"} Oct 02 09:53:00 crc kubenswrapper[4934]: I1002 09:53:00.672717 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zxq5v" podStartSLOduration=4.191465356 podStartE2EDuration="5.672683284s" podCreationTimestamp="2025-10-02 09:52:55 +0000 UTC" firstStartedPulling="2025-10-02 09:52:57.622661497 +0000 UTC m=+249.375303049" lastFinishedPulling="2025-10-02 09:52:59.103879435 +0000 UTC m=+250.856520977" observedRunningTime="2025-10-02 09:53:00.669713736 +0000 UTC m=+252.422355258" watchObservedRunningTime="2025-10-02 09:53:00.672683284 +0000 UTC m=+252.425324816" Oct 02 09:53:00 crc kubenswrapper[4934]: I1002 09:53:00.691543 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sfhmd" podStartSLOduration=3.297790291 podStartE2EDuration="4.6915192s" podCreationTimestamp="2025-10-02 09:52:56 +0000 UTC" firstStartedPulling="2025-10-02 09:52:57.628646034 +0000 UTC m=+249.381287586" lastFinishedPulling="2025-10-02 09:52:59.022374973 +0000 UTC m=+250.775016495" observedRunningTime="2025-10-02 09:53:00.688802328 +0000 UTC m=+252.441443860" watchObservedRunningTime="2025-10-02 09:53:00.6915192 +0000 UTC m=+252.444160732" Oct 02 09:53:03 crc kubenswrapper[4934]: I1002 09:53:03.905648 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4bvl5" Oct 02 09:53:03 crc kubenswrapper[4934]: I1002 09:53:03.906039 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4bvl5" Oct 02 09:53:03 crc kubenswrapper[4934]: I1002 09:53:03.952366 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4bvl5" Oct 02 09:53:04 crc kubenswrapper[4934]: I1002 09:53:04.104672 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vk4jw" Oct 02 09:53:04 crc kubenswrapper[4934]: I1002 09:53:04.105153 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vk4jw" Oct 02 09:53:04 crc kubenswrapper[4934]: I1002 09:53:04.146688 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vk4jw" Oct 02 09:53:04 crc kubenswrapper[4934]: I1002 09:53:04.749392 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vk4jw" Oct 02 09:53:04 crc kubenswrapper[4934]: I1002 09:53:04.751630 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4bvl5" Oct 02 09:53:06 crc kubenswrapper[4934]: I1002 09:53:06.307641 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zxq5v" Oct 02 09:53:06 crc kubenswrapper[4934]: I1002 09:53:06.307901 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zxq5v" Oct 02 09:53:06 crc kubenswrapper[4934]: I1002 09:53:06.343333 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zxq5v" Oct 02 09:53:06 crc kubenswrapper[4934]: I1002 09:53:06.502730 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sfhmd" Oct 02 09:53:06 crc kubenswrapper[4934]: I1002 09:53:06.502777 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sfhmd" Oct 02 09:53:06 crc kubenswrapper[4934]: I1002 09:53:06.544610 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sfhmd" Oct 02 09:53:06 crc kubenswrapper[4934]: I1002 09:53:06.738974 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sfhmd" Oct 02 09:53:06 crc kubenswrapper[4934]: I1002 09:53:06.751015 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zxq5v" Oct 02 09:54:38 crc kubenswrapper[4934]: I1002 09:54:38.439515 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:54:38 crc kubenswrapper[4934]: I1002 09:54:38.440256 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:55:02 crc kubenswrapper[4934]: I1002 09:55:02.401244 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-nmkpk"] Oct 02 09:55:02 crc kubenswrapper[4934]: I1002 09:55:02.404517 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" Oct 02 09:55:02 crc kubenswrapper[4934]: I1002 09:55:02.412699 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-nmkpk"] Oct 02 09:55:02 crc kubenswrapper[4934]: I1002 09:55:02.512519 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/955ec4d0-fead-4619-80b3-5dcd97106fa3-registry-tls\") pod \"image-registry-66df7c8f76-nmkpk\" (UID: \"955ec4d0-fead-4619-80b3-5dcd97106fa3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" Oct 02 09:55:02 crc kubenswrapper[4934]: I1002 09:55:02.512826 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/955ec4d0-fead-4619-80b3-5dcd97106fa3-bound-sa-token\") pod \"image-registry-66df7c8f76-nmkpk\" (UID: \"955ec4d0-fead-4619-80b3-5dcd97106fa3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" Oct 02 09:55:02 crc kubenswrapper[4934]: I1002 09:55:02.512973 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/955ec4d0-fead-4619-80b3-5dcd97106fa3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-nmkpk\" (UID: \"955ec4d0-fead-4619-80b3-5dcd97106fa3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" Oct 02 09:55:02 crc kubenswrapper[4934]: I1002 09:55:02.513057 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/955ec4d0-fead-4619-80b3-5dcd97106fa3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-nmkpk\" (UID: \"955ec4d0-fead-4619-80b3-5dcd97106fa3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" Oct 02 09:55:02 crc kubenswrapper[4934]: I1002 09:55:02.513288 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/955ec4d0-fead-4619-80b3-5dcd97106fa3-trusted-ca\") pod \"image-registry-66df7c8f76-nmkpk\" (UID: \"955ec4d0-fead-4619-80b3-5dcd97106fa3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" Oct 02 09:55:02 crc kubenswrapper[4934]: I1002 09:55:02.513475 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vr8bz\" (UniqueName: \"kubernetes.io/projected/955ec4d0-fead-4619-80b3-5dcd97106fa3-kube-api-access-vr8bz\") pod \"image-registry-66df7c8f76-nmkpk\" (UID: \"955ec4d0-fead-4619-80b3-5dcd97106fa3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" Oct 02 09:55:02 crc kubenswrapper[4934]: I1002 09:55:02.513596 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/955ec4d0-fead-4619-80b3-5dcd97106fa3-registry-certificates\") pod \"image-registry-66df7c8f76-nmkpk\" (UID: \"955ec4d0-fead-4619-80b3-5dcd97106fa3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" Oct 02 09:55:02 crc kubenswrapper[4934]: I1002 09:55:02.513728 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-nmkpk\" (UID: \"955ec4d0-fead-4619-80b3-5dcd97106fa3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" Oct 02 09:55:02 crc kubenswrapper[4934]: I1002 09:55:02.541287 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-nmkpk\" (UID: \"955ec4d0-fead-4619-80b3-5dcd97106fa3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" Oct 02 09:55:02 crc kubenswrapper[4934]: I1002 09:55:02.614562 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vr8bz\" (UniqueName: \"kubernetes.io/projected/955ec4d0-fead-4619-80b3-5dcd97106fa3-kube-api-access-vr8bz\") pod \"image-registry-66df7c8f76-nmkpk\" (UID: \"955ec4d0-fead-4619-80b3-5dcd97106fa3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" Oct 02 09:55:02 crc kubenswrapper[4934]: I1002 09:55:02.614629 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/955ec4d0-fead-4619-80b3-5dcd97106fa3-registry-certificates\") pod \"image-registry-66df7c8f76-nmkpk\" (UID: \"955ec4d0-fead-4619-80b3-5dcd97106fa3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" Oct 02 09:55:02 crc kubenswrapper[4934]: I1002 09:55:02.614669 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/955ec4d0-fead-4619-80b3-5dcd97106fa3-registry-tls\") pod \"image-registry-66df7c8f76-nmkpk\" (UID: \"955ec4d0-fead-4619-80b3-5dcd97106fa3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" Oct 02 09:55:02 crc kubenswrapper[4934]: I1002 09:55:02.614699 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/955ec4d0-fead-4619-80b3-5dcd97106fa3-bound-sa-token\") pod \"image-registry-66df7c8f76-nmkpk\" (UID: \"955ec4d0-fead-4619-80b3-5dcd97106fa3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" Oct 02 09:55:02 crc kubenswrapper[4934]: I1002 09:55:02.614738 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/955ec4d0-fead-4619-80b3-5dcd97106fa3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-nmkpk\" (UID: \"955ec4d0-fead-4619-80b3-5dcd97106fa3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" Oct 02 09:55:02 crc kubenswrapper[4934]: I1002 09:55:02.614760 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/955ec4d0-fead-4619-80b3-5dcd97106fa3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-nmkpk\" (UID: \"955ec4d0-fead-4619-80b3-5dcd97106fa3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" Oct 02 09:55:02 crc kubenswrapper[4934]: I1002 09:55:02.614806 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/955ec4d0-fead-4619-80b3-5dcd97106fa3-trusted-ca\") pod \"image-registry-66df7c8f76-nmkpk\" (UID: \"955ec4d0-fead-4619-80b3-5dcd97106fa3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" Oct 02 09:55:02 crc kubenswrapper[4934]: I1002 09:55:02.615717 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/955ec4d0-fead-4619-80b3-5dcd97106fa3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-nmkpk\" (UID: \"955ec4d0-fead-4619-80b3-5dcd97106fa3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" Oct 02 09:55:02 crc kubenswrapper[4934]: I1002 09:55:02.616302 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/955ec4d0-fead-4619-80b3-5dcd97106fa3-registry-certificates\") pod \"image-registry-66df7c8f76-nmkpk\" (UID: \"955ec4d0-fead-4619-80b3-5dcd97106fa3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" Oct 02 09:55:02 crc kubenswrapper[4934]: I1002 09:55:02.616560 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/955ec4d0-fead-4619-80b3-5dcd97106fa3-trusted-ca\") pod \"image-registry-66df7c8f76-nmkpk\" (UID: \"955ec4d0-fead-4619-80b3-5dcd97106fa3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" Oct 02 09:55:02 crc kubenswrapper[4934]: I1002 09:55:02.621410 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/955ec4d0-fead-4619-80b3-5dcd97106fa3-registry-tls\") pod \"image-registry-66df7c8f76-nmkpk\" (UID: \"955ec4d0-fead-4619-80b3-5dcd97106fa3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" Oct 02 09:55:02 crc kubenswrapper[4934]: I1002 09:55:02.621919 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/955ec4d0-fead-4619-80b3-5dcd97106fa3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-nmkpk\" (UID: \"955ec4d0-fead-4619-80b3-5dcd97106fa3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" Oct 02 09:55:02 crc kubenswrapper[4934]: I1002 09:55:02.632459 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/955ec4d0-fead-4619-80b3-5dcd97106fa3-bound-sa-token\") pod \"image-registry-66df7c8f76-nmkpk\" (UID: \"955ec4d0-fead-4619-80b3-5dcd97106fa3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" Oct 02 09:55:02 crc kubenswrapper[4934]: I1002 09:55:02.644857 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vr8bz\" (UniqueName: \"kubernetes.io/projected/955ec4d0-fead-4619-80b3-5dcd97106fa3-kube-api-access-vr8bz\") pod \"image-registry-66df7c8f76-nmkpk\" (UID: \"955ec4d0-fead-4619-80b3-5dcd97106fa3\") " pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" Oct 02 09:55:02 crc kubenswrapper[4934]: I1002 09:55:02.738877 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" Oct 02 09:55:02 crc kubenswrapper[4934]: I1002 09:55:02.982670 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-nmkpk"] Oct 02 09:55:02 crc kubenswrapper[4934]: W1002 09:55:02.990571 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod955ec4d0_fead_4619_80b3_5dcd97106fa3.slice/crio-cb30b8a73c111c298ca428beba6f336279bf404ab036fb6f698d0fd97982b68c WatchSource:0}: Error finding container cb30b8a73c111c298ca428beba6f336279bf404ab036fb6f698d0fd97982b68c: Status 404 returned error can't find the container with id cb30b8a73c111c298ca428beba6f336279bf404ab036fb6f698d0fd97982b68c Oct 02 09:55:03 crc kubenswrapper[4934]: I1002 09:55:03.354424 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" event={"ID":"955ec4d0-fead-4619-80b3-5dcd97106fa3","Type":"ContainerStarted","Data":"c1e972dc7f192dc61398be0b4b6150c0ea4eb37ebb618d11fb7aa8f08b2544a7"} Oct 02 09:55:03 crc kubenswrapper[4934]: I1002 09:55:03.354873 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" Oct 02 09:55:03 crc kubenswrapper[4934]: I1002 09:55:03.354899 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" event={"ID":"955ec4d0-fead-4619-80b3-5dcd97106fa3","Type":"ContainerStarted","Data":"cb30b8a73c111c298ca428beba6f336279bf404ab036fb6f698d0fd97982b68c"} Oct 02 09:55:03 crc kubenswrapper[4934]: I1002 09:55:03.385440 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" podStartSLOduration=1.385401212 podStartE2EDuration="1.385401212s" podCreationTimestamp="2025-10-02 09:55:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:55:03.376012249 +0000 UTC m=+375.128653771" watchObservedRunningTime="2025-10-02 09:55:03.385401212 +0000 UTC m=+375.138042734" Oct 02 09:55:08 crc kubenswrapper[4934]: I1002 09:55:08.439712 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:55:08 crc kubenswrapper[4934]: I1002 09:55:08.440108 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:55:22 crc kubenswrapper[4934]: I1002 09:55:22.745079 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-nmkpk" Oct 02 09:55:22 crc kubenswrapper[4934]: I1002 09:55:22.809562 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jht5p"] Oct 02 09:55:38 crc kubenswrapper[4934]: I1002 09:55:38.439982 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:55:38 crc kubenswrapper[4934]: I1002 09:55:38.440702 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:55:38 crc kubenswrapper[4934]: I1002 09:55:38.441169 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 09:55:38 crc kubenswrapper[4934]: I1002 09:55:38.442343 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0b2fff80ffe94f1c29ff579444a6c9ad1f61b9fbabc6b5775d2e9293c428ca14"} pod="openshift-machine-config-operator/machine-config-daemon-djh5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:55:38 crc kubenswrapper[4934]: I1002 09:55:38.443213 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" containerID="cri-o://0b2fff80ffe94f1c29ff579444a6c9ad1f61b9fbabc6b5775d2e9293c428ca14" gracePeriod=600 Oct 02 09:55:38 crc kubenswrapper[4934]: I1002 09:55:38.595319 4934 generic.go:334] "Generic (PLEG): container finished" podID="71db06ef-05b0-4f58-b251-b27117a8500a" containerID="0b2fff80ffe94f1c29ff579444a6c9ad1f61b9fbabc6b5775d2e9293c428ca14" exitCode=0 Oct 02 09:55:38 crc kubenswrapper[4934]: I1002 09:55:38.595706 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerDied","Data":"0b2fff80ffe94f1c29ff579444a6c9ad1f61b9fbabc6b5775d2e9293c428ca14"} Oct 02 09:55:38 crc kubenswrapper[4934]: I1002 09:55:38.595747 4934 scope.go:117] "RemoveContainer" containerID="42d6d18aae6b54a33602ec9c15ed8d69575191e5934dee425977868588d8f3af" Oct 02 09:55:39 crc kubenswrapper[4934]: I1002 09:55:39.606748 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerStarted","Data":"5d64ebbdf408852984cc94e96c4d8071b186211b51e5b9818eed5258b24a72f5"} Oct 02 09:55:47 crc kubenswrapper[4934]: I1002 09:55:47.848172 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" podUID="39e0228e-f4f6-4167-9bbc-d319374345a1" containerName="registry" containerID="cri-o://fa4e833287fc888595e5a530be3afd339fb3ef224dafb04c4032ae16485aef73" gracePeriod=30 Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.205472 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.405971 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/39e0228e-f4f6-4167-9bbc-d319374345a1-trusted-ca\") pod \"39e0228e-f4f6-4167-9bbc-d319374345a1\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.406046 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/39e0228e-f4f6-4167-9bbc-d319374345a1-ca-trust-extracted\") pod \"39e0228e-f4f6-4167-9bbc-d319374345a1\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.406094 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/39e0228e-f4f6-4167-9bbc-d319374345a1-registry-tls\") pod \"39e0228e-f4f6-4167-9bbc-d319374345a1\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.406186 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/39e0228e-f4f6-4167-9bbc-d319374345a1-registry-certificates\") pod \"39e0228e-f4f6-4167-9bbc-d319374345a1\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.406231 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/39e0228e-f4f6-4167-9bbc-d319374345a1-installation-pull-secrets\") pod \"39e0228e-f4f6-4167-9bbc-d319374345a1\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.406268 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/39e0228e-f4f6-4167-9bbc-d319374345a1-bound-sa-token\") pod \"39e0228e-f4f6-4167-9bbc-d319374345a1\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.406413 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"39e0228e-f4f6-4167-9bbc-d319374345a1\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.406459 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvmvq\" (UniqueName: \"kubernetes.io/projected/39e0228e-f4f6-4167-9bbc-d319374345a1-kube-api-access-fvmvq\") pod \"39e0228e-f4f6-4167-9bbc-d319374345a1\" (UID: \"39e0228e-f4f6-4167-9bbc-d319374345a1\") " Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.407151 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39e0228e-f4f6-4167-9bbc-d319374345a1-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "39e0228e-f4f6-4167-9bbc-d319374345a1" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.407208 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39e0228e-f4f6-4167-9bbc-d319374345a1-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "39e0228e-f4f6-4167-9bbc-d319374345a1" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.411896 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39e0228e-f4f6-4167-9bbc-d319374345a1-kube-api-access-fvmvq" (OuterVolumeSpecName: "kube-api-access-fvmvq") pod "39e0228e-f4f6-4167-9bbc-d319374345a1" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1"). InnerVolumeSpecName "kube-api-access-fvmvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.412713 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39e0228e-f4f6-4167-9bbc-d319374345a1-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "39e0228e-f4f6-4167-9bbc-d319374345a1" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.412871 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39e0228e-f4f6-4167-9bbc-d319374345a1-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "39e0228e-f4f6-4167-9bbc-d319374345a1" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.415393 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "39e0228e-f4f6-4167-9bbc-d319374345a1" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.422734 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39e0228e-f4f6-4167-9bbc-d319374345a1-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "39e0228e-f4f6-4167-9bbc-d319374345a1" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.423009 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39e0228e-f4f6-4167-9bbc-d319374345a1-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "39e0228e-f4f6-4167-9bbc-d319374345a1" (UID: "39e0228e-f4f6-4167-9bbc-d319374345a1"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.507717 4934 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/39e0228e-f4f6-4167-9bbc-d319374345a1-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.507751 4934 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/39e0228e-f4f6-4167-9bbc-d319374345a1-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.507763 4934 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/39e0228e-f4f6-4167-9bbc-d319374345a1-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.507771 4934 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/39e0228e-f4f6-4167-9bbc-d319374345a1-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.507782 4934 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/39e0228e-f4f6-4167-9bbc-d319374345a1-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.507791 4934 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/39e0228e-f4f6-4167-9bbc-d319374345a1-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.507800 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvmvq\" (UniqueName: \"kubernetes.io/projected/39e0228e-f4f6-4167-9bbc-d319374345a1-kube-api-access-fvmvq\") on node \"crc\" DevicePath \"\"" Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.668849 4934 generic.go:334] "Generic (PLEG): container finished" podID="39e0228e-f4f6-4167-9bbc-d319374345a1" containerID="fa4e833287fc888595e5a530be3afd339fb3ef224dafb04c4032ae16485aef73" exitCode=0 Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.668968 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.668968 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" event={"ID":"39e0228e-f4f6-4167-9bbc-d319374345a1","Type":"ContainerDied","Data":"fa4e833287fc888595e5a530be3afd339fb3ef224dafb04c4032ae16485aef73"} Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.669493 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-jht5p" event={"ID":"39e0228e-f4f6-4167-9bbc-d319374345a1","Type":"ContainerDied","Data":"9c6741c48da97be054129169423fe482a7880193c8c331dd0144ab1ae4dba703"} Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.669525 4934 scope.go:117] "RemoveContainer" containerID="fa4e833287fc888595e5a530be3afd339fb3ef224dafb04c4032ae16485aef73" Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.687293 4934 scope.go:117] "RemoveContainer" containerID="fa4e833287fc888595e5a530be3afd339fb3ef224dafb04c4032ae16485aef73" Oct 02 09:55:48 crc kubenswrapper[4934]: E1002 09:55:48.687820 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa4e833287fc888595e5a530be3afd339fb3ef224dafb04c4032ae16485aef73\": container with ID starting with fa4e833287fc888595e5a530be3afd339fb3ef224dafb04c4032ae16485aef73 not found: ID does not exist" containerID="fa4e833287fc888595e5a530be3afd339fb3ef224dafb04c4032ae16485aef73" Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.687884 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa4e833287fc888595e5a530be3afd339fb3ef224dafb04c4032ae16485aef73"} err="failed to get container status \"fa4e833287fc888595e5a530be3afd339fb3ef224dafb04c4032ae16485aef73\": rpc error: code = NotFound desc = could not find container \"fa4e833287fc888595e5a530be3afd339fb3ef224dafb04c4032ae16485aef73\": container with ID starting with fa4e833287fc888595e5a530be3afd339fb3ef224dafb04c4032ae16485aef73 not found: ID does not exist" Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.704411 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jht5p"] Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.708373 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-jht5p"] Oct 02 09:55:48 crc kubenswrapper[4934]: I1002 09:55:48.924114 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39e0228e-f4f6-4167-9bbc-d319374345a1" path="/var/lib/kubelet/pods/39e0228e-f4f6-4167-9bbc-d319374345a1/volumes" Oct 02 09:57:38 crc kubenswrapper[4934]: I1002 09:57:38.439981 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:57:38 crc kubenswrapper[4934]: I1002 09:57:38.440795 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:57:49 crc kubenswrapper[4934]: I1002 09:57:49.152483 4934 scope.go:117] "RemoveContainer" containerID="fcd791c407cfe2fc00e7b675bb48c2eee045ee451b731b40ef99a229ced72ae9" Oct 02 09:58:08 crc kubenswrapper[4934]: I1002 09:58:08.440041 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:58:08 crc kubenswrapper[4934]: I1002 09:58:08.440799 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:58:38 crc kubenswrapper[4934]: I1002 09:58:38.439413 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:58:38 crc kubenswrapper[4934]: I1002 09:58:38.440054 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:58:38 crc kubenswrapper[4934]: I1002 09:58:38.440122 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 09:58:38 crc kubenswrapper[4934]: I1002 09:58:38.440947 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5d64ebbdf408852984cc94e96c4d8071b186211b51e5b9818eed5258b24a72f5"} pod="openshift-machine-config-operator/machine-config-daemon-djh5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:58:38 crc kubenswrapper[4934]: I1002 09:58:38.441096 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" containerID="cri-o://5d64ebbdf408852984cc94e96c4d8071b186211b51e5b9818eed5258b24a72f5" gracePeriod=600 Oct 02 09:58:38 crc kubenswrapper[4934]: I1002 09:58:38.713839 4934 generic.go:334] "Generic (PLEG): container finished" podID="71db06ef-05b0-4f58-b251-b27117a8500a" containerID="5d64ebbdf408852984cc94e96c4d8071b186211b51e5b9818eed5258b24a72f5" exitCode=0 Oct 02 09:58:38 crc kubenswrapper[4934]: I1002 09:58:38.713914 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerDied","Data":"5d64ebbdf408852984cc94e96c4d8071b186211b51e5b9818eed5258b24a72f5"} Oct 02 09:58:38 crc kubenswrapper[4934]: I1002 09:58:38.714447 4934 scope.go:117] "RemoveContainer" containerID="0b2fff80ffe94f1c29ff579444a6c9ad1f61b9fbabc6b5775d2e9293c428ca14" Oct 02 09:58:39 crc kubenswrapper[4934]: I1002 09:58:39.727139 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerStarted","Data":"3834d18542216db82e6fdc0858cf6c8272b05593949e3231352e2c84300c22de"} Oct 02 09:58:49 crc kubenswrapper[4934]: I1002 09:58:49.186680 4934 scope.go:117] "RemoveContainer" containerID="f948029f5bf66c725b94a8a7b743870b6258a1dd1ac6482c29f587702a519a61" Oct 02 09:58:49 crc kubenswrapper[4934]: I1002 09:58:49.216143 4934 scope.go:117] "RemoveContainer" containerID="647cb41b0117107c7144477c0a14b676be611d916ed38b1dcb7d32433b07a0ee" Oct 02 09:59:51 crc kubenswrapper[4934]: I1002 09:59:51.130301 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-dd696"] Oct 02 09:59:51 crc kubenswrapper[4934]: E1002 09:59:51.131328 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39e0228e-f4f6-4167-9bbc-d319374345a1" containerName="registry" Oct 02 09:59:51 crc kubenswrapper[4934]: I1002 09:59:51.131350 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="39e0228e-f4f6-4167-9bbc-d319374345a1" containerName="registry" Oct 02 09:59:51 crc kubenswrapper[4934]: I1002 09:59:51.131535 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="39e0228e-f4f6-4167-9bbc-d319374345a1" containerName="registry" Oct 02 09:59:51 crc kubenswrapper[4934]: I1002 09:59:51.132123 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-dd696" Oct 02 09:59:51 crc kubenswrapper[4934]: I1002 09:59:51.135262 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 02 09:59:51 crc kubenswrapper[4934]: I1002 09:59:51.135900 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 02 09:59:51 crc kubenswrapper[4934]: I1002 09:59:51.136292 4934 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-5g2rw" Oct 02 09:59:51 crc kubenswrapper[4934]: I1002 09:59:51.136525 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 02 09:59:51 crc kubenswrapper[4934]: I1002 09:59:51.142558 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-dd696"] Oct 02 09:59:51 crc kubenswrapper[4934]: I1002 09:59:51.278118 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/7ce524dc-9201-41cd-8dce-78bb7d426c0f-node-mnt\") pod \"crc-storage-crc-dd696\" (UID: \"7ce524dc-9201-41cd-8dce-78bb7d426c0f\") " pod="crc-storage/crc-storage-crc-dd696" Oct 02 09:59:51 crc kubenswrapper[4934]: I1002 09:59:51.278261 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/7ce524dc-9201-41cd-8dce-78bb7d426c0f-crc-storage\") pod \"crc-storage-crc-dd696\" (UID: \"7ce524dc-9201-41cd-8dce-78bb7d426c0f\") " pod="crc-storage/crc-storage-crc-dd696" Oct 02 09:59:51 crc kubenswrapper[4934]: I1002 09:59:51.278299 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlgxs\" (UniqueName: \"kubernetes.io/projected/7ce524dc-9201-41cd-8dce-78bb7d426c0f-kube-api-access-vlgxs\") pod \"crc-storage-crc-dd696\" (UID: \"7ce524dc-9201-41cd-8dce-78bb7d426c0f\") " pod="crc-storage/crc-storage-crc-dd696" Oct 02 09:59:51 crc kubenswrapper[4934]: I1002 09:59:51.379850 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/7ce524dc-9201-41cd-8dce-78bb7d426c0f-node-mnt\") pod \"crc-storage-crc-dd696\" (UID: \"7ce524dc-9201-41cd-8dce-78bb7d426c0f\") " pod="crc-storage/crc-storage-crc-dd696" Oct 02 09:59:51 crc kubenswrapper[4934]: I1002 09:59:51.380020 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/7ce524dc-9201-41cd-8dce-78bb7d426c0f-crc-storage\") pod \"crc-storage-crc-dd696\" (UID: \"7ce524dc-9201-41cd-8dce-78bb7d426c0f\") " pod="crc-storage/crc-storage-crc-dd696" Oct 02 09:59:51 crc kubenswrapper[4934]: I1002 09:59:51.380073 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlgxs\" (UniqueName: \"kubernetes.io/projected/7ce524dc-9201-41cd-8dce-78bb7d426c0f-kube-api-access-vlgxs\") pod \"crc-storage-crc-dd696\" (UID: \"7ce524dc-9201-41cd-8dce-78bb7d426c0f\") " pod="crc-storage/crc-storage-crc-dd696" Oct 02 09:59:51 crc kubenswrapper[4934]: I1002 09:59:51.380246 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/7ce524dc-9201-41cd-8dce-78bb7d426c0f-node-mnt\") pod \"crc-storage-crc-dd696\" (UID: \"7ce524dc-9201-41cd-8dce-78bb7d426c0f\") " pod="crc-storage/crc-storage-crc-dd696" Oct 02 09:59:51 crc kubenswrapper[4934]: I1002 09:59:51.382197 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/7ce524dc-9201-41cd-8dce-78bb7d426c0f-crc-storage\") pod \"crc-storage-crc-dd696\" (UID: \"7ce524dc-9201-41cd-8dce-78bb7d426c0f\") " pod="crc-storage/crc-storage-crc-dd696" Oct 02 09:59:51 crc kubenswrapper[4934]: I1002 09:59:51.405844 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlgxs\" (UniqueName: \"kubernetes.io/projected/7ce524dc-9201-41cd-8dce-78bb7d426c0f-kube-api-access-vlgxs\") pod \"crc-storage-crc-dd696\" (UID: \"7ce524dc-9201-41cd-8dce-78bb7d426c0f\") " pod="crc-storage/crc-storage-crc-dd696" Oct 02 09:59:51 crc kubenswrapper[4934]: I1002 09:59:51.459064 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-dd696" Oct 02 09:59:51 crc kubenswrapper[4934]: I1002 09:59:51.660003 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-dd696"] Oct 02 09:59:51 crc kubenswrapper[4934]: I1002 09:59:51.668691 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 09:59:52 crc kubenswrapper[4934]: I1002 09:59:52.200883 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-dd696" event={"ID":"7ce524dc-9201-41cd-8dce-78bb7d426c0f","Type":"ContainerStarted","Data":"cb9311b118a14f14c1bd0da518bdc996b7d544e6567011158d93be6a52d7f0bd"} Oct 02 09:59:54 crc kubenswrapper[4934]: I1002 09:59:54.216964 4934 generic.go:334] "Generic (PLEG): container finished" podID="7ce524dc-9201-41cd-8dce-78bb7d426c0f" containerID="82ee58893f5546576c1c4549e0af97db7e4630a5ae1560641568690be8431fd1" exitCode=0 Oct 02 09:59:54 crc kubenswrapper[4934]: I1002 09:59:54.217060 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-dd696" event={"ID":"7ce524dc-9201-41cd-8dce-78bb7d426c0f","Type":"ContainerDied","Data":"82ee58893f5546576c1c4549e0af97db7e4630a5ae1560641568690be8431fd1"} Oct 02 09:59:55 crc kubenswrapper[4934]: I1002 09:59:55.441413 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-dd696" Oct 02 09:59:55 crc kubenswrapper[4934]: I1002 09:59:55.543353 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/7ce524dc-9201-41cd-8dce-78bb7d426c0f-crc-storage\") pod \"7ce524dc-9201-41cd-8dce-78bb7d426c0f\" (UID: \"7ce524dc-9201-41cd-8dce-78bb7d426c0f\") " Oct 02 09:59:55 crc kubenswrapper[4934]: I1002 09:59:55.543484 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlgxs\" (UniqueName: \"kubernetes.io/projected/7ce524dc-9201-41cd-8dce-78bb7d426c0f-kube-api-access-vlgxs\") pod \"7ce524dc-9201-41cd-8dce-78bb7d426c0f\" (UID: \"7ce524dc-9201-41cd-8dce-78bb7d426c0f\") " Oct 02 09:59:55 crc kubenswrapper[4934]: I1002 09:59:55.543569 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/7ce524dc-9201-41cd-8dce-78bb7d426c0f-node-mnt\") pod \"7ce524dc-9201-41cd-8dce-78bb7d426c0f\" (UID: \"7ce524dc-9201-41cd-8dce-78bb7d426c0f\") " Oct 02 09:59:55 crc kubenswrapper[4934]: I1002 09:59:55.543627 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7ce524dc-9201-41cd-8dce-78bb7d426c0f-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "7ce524dc-9201-41cd-8dce-78bb7d426c0f" (UID: "7ce524dc-9201-41cd-8dce-78bb7d426c0f"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:59:55 crc kubenswrapper[4934]: I1002 09:59:55.543920 4934 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/7ce524dc-9201-41cd-8dce-78bb7d426c0f-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:55 crc kubenswrapper[4934]: I1002 09:59:55.550542 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ce524dc-9201-41cd-8dce-78bb7d426c0f-kube-api-access-vlgxs" (OuterVolumeSpecName: "kube-api-access-vlgxs") pod "7ce524dc-9201-41cd-8dce-78bb7d426c0f" (UID: "7ce524dc-9201-41cd-8dce-78bb7d426c0f"). InnerVolumeSpecName "kube-api-access-vlgxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:59:55 crc kubenswrapper[4934]: I1002 09:59:55.565116 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ce524dc-9201-41cd-8dce-78bb7d426c0f-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "7ce524dc-9201-41cd-8dce-78bb7d426c0f" (UID: "7ce524dc-9201-41cd-8dce-78bb7d426c0f"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:59:55 crc kubenswrapper[4934]: I1002 09:59:55.645325 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlgxs\" (UniqueName: \"kubernetes.io/projected/7ce524dc-9201-41cd-8dce-78bb7d426c0f-kube-api-access-vlgxs\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:55 crc kubenswrapper[4934]: I1002 09:59:55.645365 4934 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/7ce524dc-9201-41cd-8dce-78bb7d426c0f-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 02 09:59:56 crc kubenswrapper[4934]: I1002 09:59:56.230067 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-dd696" event={"ID":"7ce524dc-9201-41cd-8dce-78bb7d426c0f","Type":"ContainerDied","Data":"cb9311b118a14f14c1bd0da518bdc996b7d544e6567011158d93be6a52d7f0bd"} Oct 02 09:59:56 crc kubenswrapper[4934]: I1002 09:59:56.230374 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb9311b118a14f14c1bd0da518bdc996b7d544e6567011158d93be6a52d7f0bd" Oct 02 09:59:56 crc kubenswrapper[4934]: I1002 09:59:56.230419 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-dd696" Oct 02 10:00:00 crc kubenswrapper[4934]: I1002 10:00:00.143007 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323320-cmf7d"] Oct 02 10:00:00 crc kubenswrapper[4934]: E1002 10:00:00.143926 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ce524dc-9201-41cd-8dce-78bb7d426c0f" containerName="storage" Oct 02 10:00:00 crc kubenswrapper[4934]: I1002 10:00:00.143946 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ce524dc-9201-41cd-8dce-78bb7d426c0f" containerName="storage" Oct 02 10:00:00 crc kubenswrapper[4934]: I1002 10:00:00.144088 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ce524dc-9201-41cd-8dce-78bb7d426c0f" containerName="storage" Oct 02 10:00:00 crc kubenswrapper[4934]: I1002 10:00:00.144748 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-cmf7d" Oct 02 10:00:00 crc kubenswrapper[4934]: I1002 10:00:00.146914 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 10:00:00 crc kubenswrapper[4934]: I1002 10:00:00.147191 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 10:00:00 crc kubenswrapper[4934]: I1002 10:00:00.151825 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323320-cmf7d"] Oct 02 10:00:00 crc kubenswrapper[4934]: I1002 10:00:00.319508 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gmgz\" (UniqueName: \"kubernetes.io/projected/c2c83244-b4c5-403e-91d0-481537b764fe-kube-api-access-7gmgz\") pod \"collect-profiles-29323320-cmf7d\" (UID: \"c2c83244-b4c5-403e-91d0-481537b764fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-cmf7d" Oct 02 10:00:00 crc kubenswrapper[4934]: I1002 10:00:00.319904 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2c83244-b4c5-403e-91d0-481537b764fe-secret-volume\") pod \"collect-profiles-29323320-cmf7d\" (UID: \"c2c83244-b4c5-403e-91d0-481537b764fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-cmf7d" Oct 02 10:00:00 crc kubenswrapper[4934]: I1002 10:00:00.319997 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2c83244-b4c5-403e-91d0-481537b764fe-config-volume\") pod \"collect-profiles-29323320-cmf7d\" (UID: \"c2c83244-b4c5-403e-91d0-481537b764fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-cmf7d" Oct 02 10:00:00 crc kubenswrapper[4934]: I1002 10:00:00.420564 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gmgz\" (UniqueName: \"kubernetes.io/projected/c2c83244-b4c5-403e-91d0-481537b764fe-kube-api-access-7gmgz\") pod \"collect-profiles-29323320-cmf7d\" (UID: \"c2c83244-b4c5-403e-91d0-481537b764fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-cmf7d" Oct 02 10:00:00 crc kubenswrapper[4934]: I1002 10:00:00.420684 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2c83244-b4c5-403e-91d0-481537b764fe-secret-volume\") pod \"collect-profiles-29323320-cmf7d\" (UID: \"c2c83244-b4c5-403e-91d0-481537b764fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-cmf7d" Oct 02 10:00:00 crc kubenswrapper[4934]: I1002 10:00:00.420720 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2c83244-b4c5-403e-91d0-481537b764fe-config-volume\") pod \"collect-profiles-29323320-cmf7d\" (UID: \"c2c83244-b4c5-403e-91d0-481537b764fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-cmf7d" Oct 02 10:00:00 crc kubenswrapper[4934]: I1002 10:00:00.421637 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2c83244-b4c5-403e-91d0-481537b764fe-config-volume\") pod \"collect-profiles-29323320-cmf7d\" (UID: \"c2c83244-b4c5-403e-91d0-481537b764fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-cmf7d" Oct 02 10:00:00 crc kubenswrapper[4934]: I1002 10:00:00.426161 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2c83244-b4c5-403e-91d0-481537b764fe-secret-volume\") pod \"collect-profiles-29323320-cmf7d\" (UID: \"c2c83244-b4c5-403e-91d0-481537b764fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-cmf7d" Oct 02 10:00:00 crc kubenswrapper[4934]: I1002 10:00:00.445141 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gmgz\" (UniqueName: \"kubernetes.io/projected/c2c83244-b4c5-403e-91d0-481537b764fe-kube-api-access-7gmgz\") pod \"collect-profiles-29323320-cmf7d\" (UID: \"c2c83244-b4c5-403e-91d0-481537b764fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-cmf7d" Oct 02 10:00:00 crc kubenswrapper[4934]: I1002 10:00:00.472073 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-cmf7d" Oct 02 10:00:00 crc kubenswrapper[4934]: I1002 10:00:00.905065 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323320-cmf7d"] Oct 02 10:00:01 crc kubenswrapper[4934]: I1002 10:00:01.255653 4934 generic.go:334] "Generic (PLEG): container finished" podID="c2c83244-b4c5-403e-91d0-481537b764fe" containerID="09eb614893d7c26c569495d5413034f272a564e4ad165a4297729f1221b50851" exitCode=0 Oct 02 10:00:01 crc kubenswrapper[4934]: I1002 10:00:01.255698 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-cmf7d" event={"ID":"c2c83244-b4c5-403e-91d0-481537b764fe","Type":"ContainerDied","Data":"09eb614893d7c26c569495d5413034f272a564e4ad165a4297729f1221b50851"} Oct 02 10:00:01 crc kubenswrapper[4934]: I1002 10:00:01.255726 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-cmf7d" event={"ID":"c2c83244-b4c5-403e-91d0-481537b764fe","Type":"ContainerStarted","Data":"e6cb5e84810fc39b7fbcf37a7f9ed838986b45ebb35ce63821c766e2d39fc17d"} Oct 02 10:00:02 crc kubenswrapper[4934]: I1002 10:00:02.489502 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-cmf7d" Oct 02 10:00:02 crc kubenswrapper[4934]: I1002 10:00:02.551687 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2c83244-b4c5-403e-91d0-481537b764fe-secret-volume\") pod \"c2c83244-b4c5-403e-91d0-481537b764fe\" (UID: \"c2c83244-b4c5-403e-91d0-481537b764fe\") " Oct 02 10:00:02 crc kubenswrapper[4934]: I1002 10:00:02.551738 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2c83244-b4c5-403e-91d0-481537b764fe-config-volume\") pod \"c2c83244-b4c5-403e-91d0-481537b764fe\" (UID: \"c2c83244-b4c5-403e-91d0-481537b764fe\") " Oct 02 10:00:02 crc kubenswrapper[4934]: I1002 10:00:02.551767 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gmgz\" (UniqueName: \"kubernetes.io/projected/c2c83244-b4c5-403e-91d0-481537b764fe-kube-api-access-7gmgz\") pod \"c2c83244-b4c5-403e-91d0-481537b764fe\" (UID: \"c2c83244-b4c5-403e-91d0-481537b764fe\") " Oct 02 10:00:02 crc kubenswrapper[4934]: I1002 10:00:02.552643 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2c83244-b4c5-403e-91d0-481537b764fe-config-volume" (OuterVolumeSpecName: "config-volume") pod "c2c83244-b4c5-403e-91d0-481537b764fe" (UID: "c2c83244-b4c5-403e-91d0-481537b764fe"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:02 crc kubenswrapper[4934]: I1002 10:00:02.557301 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2c83244-b4c5-403e-91d0-481537b764fe-kube-api-access-7gmgz" (OuterVolumeSpecName: "kube-api-access-7gmgz") pod "c2c83244-b4c5-403e-91d0-481537b764fe" (UID: "c2c83244-b4c5-403e-91d0-481537b764fe"). InnerVolumeSpecName "kube-api-access-7gmgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:00:02 crc kubenswrapper[4934]: I1002 10:00:02.567900 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2c83244-b4c5-403e-91d0-481537b764fe-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c2c83244-b4c5-403e-91d0-481537b764fe" (UID: "c2c83244-b4c5-403e-91d0-481537b764fe"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:02 crc kubenswrapper[4934]: I1002 10:00:02.652684 4934 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2c83244-b4c5-403e-91d0-481537b764fe-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:02 crc kubenswrapper[4934]: I1002 10:00:02.652711 4934 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2c83244-b4c5-403e-91d0-481537b764fe-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:02 crc kubenswrapper[4934]: I1002 10:00:02.652721 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gmgz\" (UniqueName: \"kubernetes.io/projected/c2c83244-b4c5-403e-91d0-481537b764fe-kube-api-access-7gmgz\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:02 crc kubenswrapper[4934]: I1002 10:00:02.946820 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl"] Oct 02 10:00:02 crc kubenswrapper[4934]: E1002 10:00:02.947095 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2c83244-b4c5-403e-91d0-481537b764fe" containerName="collect-profiles" Oct 02 10:00:02 crc kubenswrapper[4934]: I1002 10:00:02.947109 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2c83244-b4c5-403e-91d0-481537b764fe" containerName="collect-profiles" Oct 02 10:00:02 crc kubenswrapper[4934]: I1002 10:00:02.947217 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2c83244-b4c5-403e-91d0-481537b764fe" containerName="collect-profiles" Oct 02 10:00:02 crc kubenswrapper[4934]: I1002 10:00:02.948089 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl" Oct 02 10:00:02 crc kubenswrapper[4934]: I1002 10:00:02.950302 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 02 10:00:02 crc kubenswrapper[4934]: I1002 10:00:02.955352 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rlnr\" (UniqueName: \"kubernetes.io/projected/89aa6f88-b397-4574-8ad9-6c855a2c8511-kube-api-access-9rlnr\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl\" (UID: \"89aa6f88-b397-4574-8ad9-6c855a2c8511\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl" Oct 02 10:00:02 crc kubenswrapper[4934]: I1002 10:00:02.955411 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/89aa6f88-b397-4574-8ad9-6c855a2c8511-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl\" (UID: \"89aa6f88-b397-4574-8ad9-6c855a2c8511\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl" Oct 02 10:00:02 crc kubenswrapper[4934]: I1002 10:00:02.955553 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/89aa6f88-b397-4574-8ad9-6c855a2c8511-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl\" (UID: \"89aa6f88-b397-4574-8ad9-6c855a2c8511\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl" Oct 02 10:00:02 crc kubenswrapper[4934]: I1002 10:00:02.957103 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl"] Oct 02 10:00:03 crc kubenswrapper[4934]: I1002 10:00:03.056629 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/89aa6f88-b397-4574-8ad9-6c855a2c8511-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl\" (UID: \"89aa6f88-b397-4574-8ad9-6c855a2c8511\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl" Oct 02 10:00:03 crc kubenswrapper[4934]: I1002 10:00:03.056944 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/89aa6f88-b397-4574-8ad9-6c855a2c8511-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl\" (UID: \"89aa6f88-b397-4574-8ad9-6c855a2c8511\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl" Oct 02 10:00:03 crc kubenswrapper[4934]: I1002 10:00:03.057076 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rlnr\" (UniqueName: \"kubernetes.io/projected/89aa6f88-b397-4574-8ad9-6c855a2c8511-kube-api-access-9rlnr\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl\" (UID: \"89aa6f88-b397-4574-8ad9-6c855a2c8511\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl" Oct 02 10:00:03 crc kubenswrapper[4934]: I1002 10:00:03.057270 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/89aa6f88-b397-4574-8ad9-6c855a2c8511-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl\" (UID: \"89aa6f88-b397-4574-8ad9-6c855a2c8511\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl" Oct 02 10:00:03 crc kubenswrapper[4934]: I1002 10:00:03.057713 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/89aa6f88-b397-4574-8ad9-6c855a2c8511-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl\" (UID: \"89aa6f88-b397-4574-8ad9-6c855a2c8511\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl" Oct 02 10:00:03 crc kubenswrapper[4934]: I1002 10:00:03.074599 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rlnr\" (UniqueName: \"kubernetes.io/projected/89aa6f88-b397-4574-8ad9-6c855a2c8511-kube-api-access-9rlnr\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl\" (UID: \"89aa6f88-b397-4574-8ad9-6c855a2c8511\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl" Oct 02 10:00:03 crc kubenswrapper[4934]: I1002 10:00:03.264899 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl" Oct 02 10:00:03 crc kubenswrapper[4934]: I1002 10:00:03.277405 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-cmf7d" event={"ID":"c2c83244-b4c5-403e-91d0-481537b764fe","Type":"ContainerDied","Data":"e6cb5e84810fc39b7fbcf37a7f9ed838986b45ebb35ce63821c766e2d39fc17d"} Oct 02 10:00:03 crc kubenswrapper[4934]: I1002 10:00:03.277461 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6cb5e84810fc39b7fbcf37a7f9ed838986b45ebb35ce63821c766e2d39fc17d" Oct 02 10:00:03 crc kubenswrapper[4934]: I1002 10:00:03.277498 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-cmf7d" Oct 02 10:00:03 crc kubenswrapper[4934]: I1002 10:00:03.525242 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl"] Oct 02 10:00:04 crc kubenswrapper[4934]: I1002 10:00:04.285424 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl" event={"ID":"89aa6f88-b397-4574-8ad9-6c855a2c8511","Type":"ContainerStarted","Data":"6cac11dd9f0389e3cd7ff7ab593e891b5020dab93c4c6e1891b2bf8b5d49a922"} Oct 02 10:00:04 crc kubenswrapper[4934]: I1002 10:00:04.285791 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl" event={"ID":"89aa6f88-b397-4574-8ad9-6c855a2c8511","Type":"ContainerStarted","Data":"b7a7213977b6eb6117a3c80ce003f759674f578dc92e582fca1435e58eae077a"} Oct 02 10:00:05 crc kubenswrapper[4934]: I1002 10:00:05.290238 4934 generic.go:334] "Generic (PLEG): container finished" podID="89aa6f88-b397-4574-8ad9-6c855a2c8511" containerID="6cac11dd9f0389e3cd7ff7ab593e891b5020dab93c4c6e1891b2bf8b5d49a922" exitCode=0 Oct 02 10:00:05 crc kubenswrapper[4934]: I1002 10:00:05.290324 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl" event={"ID":"89aa6f88-b397-4574-8ad9-6c855a2c8511","Type":"ContainerDied","Data":"6cac11dd9f0389e3cd7ff7ab593e891b5020dab93c4c6e1891b2bf8b5d49a922"} Oct 02 10:00:07 crc kubenswrapper[4934]: I1002 10:00:07.302484 4934 generic.go:334] "Generic (PLEG): container finished" podID="89aa6f88-b397-4574-8ad9-6c855a2c8511" containerID="3a069adc3d260491e5510fe14b09cb30e4657a4de1a285d161349afde61d837b" exitCode=0 Oct 02 10:00:07 crc kubenswrapper[4934]: I1002 10:00:07.302548 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl" event={"ID":"89aa6f88-b397-4574-8ad9-6c855a2c8511","Type":"ContainerDied","Data":"3a069adc3d260491e5510fe14b09cb30e4657a4de1a285d161349afde61d837b"} Oct 02 10:00:07 crc kubenswrapper[4934]: I1002 10:00:07.819989 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-th4v6"] Oct 02 10:00:07 crc kubenswrapper[4934]: I1002 10:00:07.820516 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="ovn-controller" containerID="cri-o://4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb" gracePeriod=30 Oct 02 10:00:07 crc kubenswrapper[4934]: I1002 10:00:07.820972 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="kube-rbac-proxy-node" containerID="cri-o://e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b" gracePeriod=30 Oct 02 10:00:07 crc kubenswrapper[4934]: I1002 10:00:07.820993 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9" gracePeriod=30 Oct 02 10:00:07 crc kubenswrapper[4934]: I1002 10:00:07.821055 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="ovn-acl-logging" containerID="cri-o://e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019" gracePeriod=30 Oct 02 10:00:07 crc kubenswrapper[4934]: I1002 10:00:07.821224 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="sbdb" containerID="cri-o://2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c" gracePeriod=30 Oct 02 10:00:07 crc kubenswrapper[4934]: I1002 10:00:07.821370 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="nbdb" containerID="cri-o://9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea" gracePeriod=30 Oct 02 10:00:07 crc kubenswrapper[4934]: I1002 10:00:07.821422 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="northd" containerID="cri-o://c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8" gracePeriod=30 Oct 02 10:00:07 crc kubenswrapper[4934]: I1002 10:00:07.871507 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="ovnkube-controller" containerID="cri-o://354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66" gracePeriod=30 Oct 02 10:00:07 crc kubenswrapper[4934]: E1002 10:00:07.887915 4934 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d4841b5_0469_461e_875c_25b9fe848141.slice/crio-conmon-e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019.scope\": RecentStats: unable to find data in memory cache]" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.190348 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-th4v6_1d4841b5-0469-461e-875c-25b9fe848141/ovnkube-controller/3.log" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.192216 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-th4v6_1d4841b5-0469-461e-875c-25b9fe848141/ovn-acl-logging/0.log" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.192870 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-th4v6_1d4841b5-0469-461e-875c-25b9fe848141/ovn-controller/0.log" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.193390 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.242030 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-kcgtd"] Oct 02 10:00:08 crc kubenswrapper[4934]: E1002 10:00:08.242330 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="ovnkube-controller" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.242356 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="ovnkube-controller" Oct 02 10:00:08 crc kubenswrapper[4934]: E1002 10:00:08.242376 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="ovn-controller" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.242388 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="ovn-controller" Oct 02 10:00:08 crc kubenswrapper[4934]: E1002 10:00:08.242406 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="kube-rbac-proxy-node" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.242418 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="kube-rbac-proxy-node" Oct 02 10:00:08 crc kubenswrapper[4934]: E1002 10:00:08.242438 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="nbdb" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.242450 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="nbdb" Oct 02 10:00:08 crc kubenswrapper[4934]: E1002 10:00:08.242469 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="ovn-acl-logging" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.242481 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="ovn-acl-logging" Oct 02 10:00:08 crc kubenswrapper[4934]: E1002 10:00:08.242496 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="ovnkube-controller" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.242508 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="ovnkube-controller" Oct 02 10:00:08 crc kubenswrapper[4934]: E1002 10:00:08.242523 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="kubecfg-setup" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.242535 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="kubecfg-setup" Oct 02 10:00:08 crc kubenswrapper[4934]: E1002 10:00:08.242552 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.242564 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 10:00:08 crc kubenswrapper[4934]: E1002 10:00:08.242685 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="northd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.242702 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="northd" Oct 02 10:00:08 crc kubenswrapper[4934]: E1002 10:00:08.242724 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="ovnkube-controller" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.242739 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="ovnkube-controller" Oct 02 10:00:08 crc kubenswrapper[4934]: E1002 10:00:08.242757 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="sbdb" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.242769 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="sbdb" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.242930 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="ovnkube-controller" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.242946 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="ovnkube-controller" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.242961 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.242979 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="ovn-controller" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.242997 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="ovnkube-controller" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.243012 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="ovn-acl-logging" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.243031 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="northd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.243046 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="ovnkube-controller" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.243067 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="sbdb" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.243081 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="kube-rbac-proxy-node" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.243098 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="nbdb" Oct 02 10:00:08 crc kubenswrapper[4934]: E1002 10:00:08.243258 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="ovnkube-controller" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.243272 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="ovnkube-controller" Oct 02 10:00:08 crc kubenswrapper[4934]: E1002 10:00:08.243296 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="ovnkube-controller" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.243310 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="ovnkube-controller" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.243473 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d4841b5-0469-461e-875c-25b9fe848141" containerName="ovnkube-controller" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.246520 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.310859 4934 generic.go:334] "Generic (PLEG): container finished" podID="89aa6f88-b397-4574-8ad9-6c855a2c8511" containerID="08a4999572374f1087b705f2819decfcb500e963f7511b72e0e9d486f59fb501" exitCode=0 Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.310940 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl" event={"ID":"89aa6f88-b397-4574-8ad9-6c855a2c8511","Type":"ContainerDied","Data":"08a4999572374f1087b705f2819decfcb500e963f7511b72e0e9d486f59fb501"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.313470 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-th4v6_1d4841b5-0469-461e-875c-25b9fe848141/ovnkube-controller/3.log" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.316183 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-th4v6_1d4841b5-0469-461e-875c-25b9fe848141/ovn-acl-logging/0.log" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.316696 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-th4v6_1d4841b5-0469-461e-875c-25b9fe848141/ovn-controller/0.log" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.317276 4934 generic.go:334] "Generic (PLEG): container finished" podID="1d4841b5-0469-461e-875c-25b9fe848141" containerID="354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66" exitCode=0 Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.317343 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.317369 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" event={"ID":"1d4841b5-0469-461e-875c-25b9fe848141","Type":"ContainerDied","Data":"354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.317424 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" event={"ID":"1d4841b5-0469-461e-875c-25b9fe848141","Type":"ContainerDied","Data":"2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.317464 4934 scope.go:117] "RemoveContainer" containerID="354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.317342 4934 generic.go:334] "Generic (PLEG): container finished" podID="1d4841b5-0469-461e-875c-25b9fe848141" containerID="2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c" exitCode=0 Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.317794 4934 generic.go:334] "Generic (PLEG): container finished" podID="1d4841b5-0469-461e-875c-25b9fe848141" containerID="9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea" exitCode=0 Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.317811 4934 generic.go:334] "Generic (PLEG): container finished" podID="1d4841b5-0469-461e-875c-25b9fe848141" containerID="c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8" exitCode=0 Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.317821 4934 generic.go:334] "Generic (PLEG): container finished" podID="1d4841b5-0469-461e-875c-25b9fe848141" containerID="389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9" exitCode=0 Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.317830 4934 generic.go:334] "Generic (PLEG): container finished" podID="1d4841b5-0469-461e-875c-25b9fe848141" containerID="e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b" exitCode=0 Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.317841 4934 generic.go:334] "Generic (PLEG): container finished" podID="1d4841b5-0469-461e-875c-25b9fe848141" containerID="e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019" exitCode=143 Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.317850 4934 generic.go:334] "Generic (PLEG): container finished" podID="1d4841b5-0469-461e-875c-25b9fe848141" containerID="4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb" exitCode=143 Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.317850 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" event={"ID":"1d4841b5-0469-461e-875c-25b9fe848141","Type":"ContainerDied","Data":"9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.317894 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" event={"ID":"1d4841b5-0469-461e-875c-25b9fe848141","Type":"ContainerDied","Data":"c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.317909 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" event={"ID":"1d4841b5-0469-461e-875c-25b9fe848141","Type":"ContainerDied","Data":"389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.317924 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" event={"ID":"1d4841b5-0469-461e-875c-25b9fe848141","Type":"ContainerDied","Data":"e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.317938 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.317950 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.317957 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.317966 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.317973 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.317979 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.317986 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.317992 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.317999 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318009 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" event={"ID":"1d4841b5-0469-461e-875c-25b9fe848141","Type":"ContainerDied","Data":"e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318021 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318029 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318036 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318042 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318049 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318056 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318063 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318069 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318076 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318083 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318092 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" event={"ID":"1d4841b5-0469-461e-875c-25b9fe848141","Type":"ContainerDied","Data":"4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318103 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318112 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318118 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318125 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318133 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318140 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318146 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318153 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318160 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318166 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318176 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-th4v6" event={"ID":"1d4841b5-0469-461e-875c-25b9fe848141","Type":"ContainerDied","Data":"bcb32bcd13898c8fdf7e282c3b8794230b382612943869b10216ed0361924b16"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318186 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318193 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318200 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318207 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318214 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318220 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318227 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318234 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318241 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.318248 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.320354 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-79fxg_73968247-67dd-48cc-88a1-64afac657412/kube-multus/1.log" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.323969 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-79fxg_73968247-67dd-48cc-88a1-64afac657412/kube-multus/0.log" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.324020 4934 generic.go:334] "Generic (PLEG): container finished" podID="73968247-67dd-48cc-88a1-64afac657412" containerID="84da7560cede71d06bd39eeee26703d00479335105cce31a29648fbd36512917" exitCode=2 Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.324058 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-79fxg" event={"ID":"73968247-67dd-48cc-88a1-64afac657412","Type":"ContainerDied","Data":"84da7560cede71d06bd39eeee26703d00479335105cce31a29648fbd36512917"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.324086 4934 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65"} Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.324715 4934 scope.go:117] "RemoveContainer" containerID="84da7560cede71d06bd39eeee26703d00479335105cce31a29648fbd36512917" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.334673 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-slash\") pod \"1d4841b5-0469-461e-875c-25b9fe848141\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.334743 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-run-netns\") pod \"1d4841b5-0469-461e-875c-25b9fe848141\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.334784 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-kubelet\") pod \"1d4841b5-0469-461e-875c-25b9fe848141\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.334833 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-run-ovn-kubernetes\") pod \"1d4841b5-0469-461e-875c-25b9fe848141\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.334865 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-node-log\") pod \"1d4841b5-0469-461e-875c-25b9fe848141\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.334911 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mgrl\" (UniqueName: \"kubernetes.io/projected/1d4841b5-0469-461e-875c-25b9fe848141-kube-api-access-4mgrl\") pod \"1d4841b5-0469-461e-875c-25b9fe848141\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.334944 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-var-lib-cni-networks-ovn-kubernetes\") pod \"1d4841b5-0469-461e-875c-25b9fe848141\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.334983 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-run-openvswitch\") pod \"1d4841b5-0469-461e-875c-25b9fe848141\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.335029 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-var-lib-openvswitch\") pod \"1d4841b5-0469-461e-875c-25b9fe848141\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.335062 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-cni-netd\") pod \"1d4841b5-0469-461e-875c-25b9fe848141\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.335092 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-run-systemd\") pod \"1d4841b5-0469-461e-875c-25b9fe848141\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.335123 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-log-socket\") pod \"1d4841b5-0469-461e-875c-25b9fe848141\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.335167 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-etc-openvswitch\") pod \"1d4841b5-0469-461e-875c-25b9fe848141\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.335201 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1d4841b5-0469-461e-875c-25b9fe848141-env-overrides\") pod \"1d4841b5-0469-461e-875c-25b9fe848141\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.335240 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-systemd-units\") pod \"1d4841b5-0469-461e-875c-25b9fe848141\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.335270 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1d4841b5-0469-461e-875c-25b9fe848141-ovnkube-config\") pod \"1d4841b5-0469-461e-875c-25b9fe848141\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.335298 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-cni-bin\") pod \"1d4841b5-0469-461e-875c-25b9fe848141\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.335346 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-run-ovn\") pod \"1d4841b5-0469-461e-875c-25b9fe848141\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.335391 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1d4841b5-0469-461e-875c-25b9fe848141-ovnkube-script-lib\") pod \"1d4841b5-0469-461e-875c-25b9fe848141\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.335420 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1d4841b5-0469-461e-875c-25b9fe848141-ovn-node-metrics-cert\") pod \"1d4841b5-0469-461e-875c-25b9fe848141\" (UID: \"1d4841b5-0469-461e-875c-25b9fe848141\") " Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.335724 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "1d4841b5-0469-461e-875c-25b9fe848141" (UID: "1d4841b5-0469-461e-875c-25b9fe848141"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.335777 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-slash" (OuterVolumeSpecName: "host-slash") pod "1d4841b5-0469-461e-875c-25b9fe848141" (UID: "1d4841b5-0469-461e-875c-25b9fe848141"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.335813 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "1d4841b5-0469-461e-875c-25b9fe848141" (UID: "1d4841b5-0469-461e-875c-25b9fe848141"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.335846 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "1d4841b5-0469-461e-875c-25b9fe848141" (UID: "1d4841b5-0469-461e-875c-25b9fe848141"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.335875 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "1d4841b5-0469-461e-875c-25b9fe848141" (UID: "1d4841b5-0469-461e-875c-25b9fe848141"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.335905 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-node-log" (OuterVolumeSpecName: "node-log") pod "1d4841b5-0469-461e-875c-25b9fe848141" (UID: "1d4841b5-0469-461e-875c-25b9fe848141"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.338967 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "1d4841b5-0469-461e-875c-25b9fe848141" (UID: "1d4841b5-0469-461e-875c-25b9fe848141"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.338980 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "1d4841b5-0469-461e-875c-25b9fe848141" (UID: "1d4841b5-0469-461e-875c-25b9fe848141"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.339023 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "1d4841b5-0469-461e-875c-25b9fe848141" (UID: "1d4841b5-0469-461e-875c-25b9fe848141"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.339003 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "1d4841b5-0469-461e-875c-25b9fe848141" (UID: "1d4841b5-0469-461e-875c-25b9fe848141"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.339045 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-log-socket" (OuterVolumeSpecName: "log-socket") pod "1d4841b5-0469-461e-875c-25b9fe848141" (UID: "1d4841b5-0469-461e-875c-25b9fe848141"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.339062 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "1d4841b5-0469-461e-875c-25b9fe848141" (UID: "1d4841b5-0469-461e-875c-25b9fe848141"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.339106 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "1d4841b5-0469-461e-875c-25b9fe848141" (UID: "1d4841b5-0469-461e-875c-25b9fe848141"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.339154 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "1d4841b5-0469-461e-875c-25b9fe848141" (UID: "1d4841b5-0469-461e-875c-25b9fe848141"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.339529 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d4841b5-0469-461e-875c-25b9fe848141-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "1d4841b5-0469-461e-875c-25b9fe848141" (UID: "1d4841b5-0469-461e-875c-25b9fe848141"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.339937 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d4841b5-0469-461e-875c-25b9fe848141-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "1d4841b5-0469-461e-875c-25b9fe848141" (UID: "1d4841b5-0469-461e-875c-25b9fe848141"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.340068 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d4841b5-0469-461e-875c-25b9fe848141-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "1d4841b5-0469-461e-875c-25b9fe848141" (UID: "1d4841b5-0469-461e-875c-25b9fe848141"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.343859 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-host-run-ovn-kubernetes\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.343949 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-529c6\" (UniqueName: \"kubernetes.io/projected/4bffef6b-7a79-4c97-a874-686223e44047-kube-api-access-529c6\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.343986 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-run-systemd\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.344014 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-etc-openvswitch\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.344042 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4bffef6b-7a79-4c97-a874-686223e44047-env-overrides\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.344078 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-var-lib-openvswitch\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.344105 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4bffef6b-7a79-4c97-a874-686223e44047-ovnkube-script-lib\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.344233 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-host-run-netns\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.344301 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-run-ovn\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.344369 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-systemd-units\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.344441 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.344476 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-node-log\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.344538 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-host-cni-bin\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.344663 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-host-kubelet\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.344807 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d4841b5-0469-461e-875c-25b9fe848141-kube-api-access-4mgrl" (OuterVolumeSpecName: "kube-api-access-4mgrl") pod "1d4841b5-0469-461e-875c-25b9fe848141" (UID: "1d4841b5-0469-461e-875c-25b9fe848141"). InnerVolumeSpecName "kube-api-access-4mgrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.344942 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d4841b5-0469-461e-875c-25b9fe848141-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "1d4841b5-0469-461e-875c-25b9fe848141" (UID: "1d4841b5-0469-461e-875c-25b9fe848141"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.345285 4934 scope.go:117] "RemoveContainer" containerID="358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.349255 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-host-cni-netd\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.349342 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-log-socket\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.349412 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-host-slash\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.349438 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-run-openvswitch\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.349462 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4bffef6b-7a79-4c97-a874-686223e44047-ovnkube-config\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.349480 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4bffef6b-7a79-4c97-a874-686223e44047-ovn-node-metrics-cert\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.349594 4934 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.349609 4934 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1d4841b5-0469-461e-875c-25b9fe848141-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.349622 4934 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.349633 4934 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1d4841b5-0469-461e-875c-25b9fe848141-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.349644 4934 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.349657 4934 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.349669 4934 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1d4841b5-0469-461e-875c-25b9fe848141-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.349683 4934 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1d4841b5-0469-461e-875c-25b9fe848141-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.349696 4934 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-slash\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.349705 4934 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.349718 4934 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.349729 4934 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.349740 4934 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-node-log\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.349751 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mgrl\" (UniqueName: \"kubernetes.io/projected/1d4841b5-0469-461e-875c-25b9fe848141-kube-api-access-4mgrl\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.349761 4934 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.349771 4934 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.349780 4934 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.349787 4934 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.349796 4934 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-log-socket\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.357826 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "1d4841b5-0469-461e-875c-25b9fe848141" (UID: "1d4841b5-0469-461e-875c-25b9fe848141"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.375179 4934 scope.go:117] "RemoveContainer" containerID="2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.389677 4934 scope.go:117] "RemoveContainer" containerID="9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.414216 4934 scope.go:117] "RemoveContainer" containerID="c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.445385 4934 scope.go:117] "RemoveContainer" containerID="389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.451333 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-host-cni-bin\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.451454 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-host-kubelet\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.451502 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-host-cni-netd\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.451459 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-host-cni-bin\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.451543 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-host-cni-netd\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.451551 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-log-socket\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.451638 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-log-socket\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.451659 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-host-slash\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.451502 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-host-kubelet\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.451687 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-run-openvswitch\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.451718 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-host-slash\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.451730 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4bffef6b-7a79-4c97-a874-686223e44047-ovnkube-config\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.451761 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-run-openvswitch\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.451766 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4bffef6b-7a79-4c97-a874-686223e44047-ovn-node-metrics-cert\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.451819 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-host-run-ovn-kubernetes\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.451855 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-run-systemd\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.451887 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-529c6\" (UniqueName: \"kubernetes.io/projected/4bffef6b-7a79-4c97-a874-686223e44047-kube-api-access-529c6\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.451917 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-etc-openvswitch\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.451924 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-host-run-ovn-kubernetes\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.451949 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4bffef6b-7a79-4c97-a874-686223e44047-env-overrides\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.451981 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4bffef6b-7a79-4c97-a874-686223e44047-ovnkube-script-lib\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.451989 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-etc-openvswitch\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.452013 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-var-lib-openvswitch\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.452050 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-host-run-netns\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.452087 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-run-ovn\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.452062 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-run-systemd\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.452132 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-systemd-units\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.452183 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-var-lib-openvswitch\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.452204 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.452233 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-run-ovn\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.452241 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-node-log\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.452268 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-host-run-netns\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.452331 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.452330 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-systemd-units\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.452330 4934 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1d4841b5-0469-461e-875c-25b9fe848141-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.452392 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4bffef6b-7a79-4c97-a874-686223e44047-node-log\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.452527 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4bffef6b-7a79-4c97-a874-686223e44047-env-overrides\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.457121 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4bffef6b-7a79-4c97-a874-686223e44047-ovnkube-config\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.457227 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4bffef6b-7a79-4c97-a874-686223e44047-ovnkube-script-lib\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.458700 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4bffef6b-7a79-4c97-a874-686223e44047-ovn-node-metrics-cert\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.461837 4934 scope.go:117] "RemoveContainer" containerID="e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.471405 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-529c6\" (UniqueName: \"kubernetes.io/projected/4bffef6b-7a79-4c97-a874-686223e44047-kube-api-access-529c6\") pod \"ovnkube-node-kcgtd\" (UID: \"4bffef6b-7a79-4c97-a874-686223e44047\") " pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.477778 4934 scope.go:117] "RemoveContainer" containerID="e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.492622 4934 scope.go:117] "RemoveContainer" containerID="4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.510666 4934 scope.go:117] "RemoveContainer" containerID="46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.531855 4934 scope.go:117] "RemoveContainer" containerID="354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66" Oct 02 10:00:08 crc kubenswrapper[4934]: E1002 10:00:08.532310 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66\": container with ID starting with 354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66 not found: ID does not exist" containerID="354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.532348 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66"} err="failed to get container status \"354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66\": rpc error: code = NotFound desc = could not find container \"354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66\": container with ID starting with 354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66 not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.532375 4934 scope.go:117] "RemoveContainer" containerID="358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe" Oct 02 10:00:08 crc kubenswrapper[4934]: E1002 10:00:08.532615 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe\": container with ID starting with 358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe not found: ID does not exist" containerID="358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.532638 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe"} err="failed to get container status \"358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe\": rpc error: code = NotFound desc = could not find container \"358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe\": container with ID starting with 358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.532651 4934 scope.go:117] "RemoveContainer" containerID="2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c" Oct 02 10:00:08 crc kubenswrapper[4934]: E1002 10:00:08.532892 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\": container with ID starting with 2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c not found: ID does not exist" containerID="2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.532967 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c"} err="failed to get container status \"2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\": rpc error: code = NotFound desc = could not find container \"2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\": container with ID starting with 2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.532998 4934 scope.go:117] "RemoveContainer" containerID="9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea" Oct 02 10:00:08 crc kubenswrapper[4934]: E1002 10:00:08.534300 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\": container with ID starting with 9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea not found: ID does not exist" containerID="9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.534345 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea"} err="failed to get container status \"9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\": rpc error: code = NotFound desc = could not find container \"9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\": container with ID starting with 9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.534359 4934 scope.go:117] "RemoveContainer" containerID="c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8" Oct 02 10:00:08 crc kubenswrapper[4934]: E1002 10:00:08.535655 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\": container with ID starting with c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8 not found: ID does not exist" containerID="c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.535683 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8"} err="failed to get container status \"c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\": rpc error: code = NotFound desc = could not find container \"c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\": container with ID starting with c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8 not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.535700 4934 scope.go:117] "RemoveContainer" containerID="389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9" Oct 02 10:00:08 crc kubenswrapper[4934]: E1002 10:00:08.535982 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\": container with ID starting with 389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9 not found: ID does not exist" containerID="389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.536014 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9"} err="failed to get container status \"389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\": rpc error: code = NotFound desc = could not find container \"389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\": container with ID starting with 389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9 not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.536034 4934 scope.go:117] "RemoveContainer" containerID="e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b" Oct 02 10:00:08 crc kubenswrapper[4934]: E1002 10:00:08.536299 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\": container with ID starting with e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b not found: ID does not exist" containerID="e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.536331 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b"} err="failed to get container status \"e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\": rpc error: code = NotFound desc = could not find container \"e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\": container with ID starting with e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.536352 4934 scope.go:117] "RemoveContainer" containerID="e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019" Oct 02 10:00:08 crc kubenswrapper[4934]: E1002 10:00:08.536725 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\": container with ID starting with e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019 not found: ID does not exist" containerID="e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.536800 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019"} err="failed to get container status \"e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\": rpc error: code = NotFound desc = could not find container \"e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\": container with ID starting with e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019 not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.536819 4934 scope.go:117] "RemoveContainer" containerID="4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb" Oct 02 10:00:08 crc kubenswrapper[4934]: E1002 10:00:08.537106 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\": container with ID starting with 4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb not found: ID does not exist" containerID="4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.537135 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb"} err="failed to get container status \"4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\": rpc error: code = NotFound desc = could not find container \"4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\": container with ID starting with 4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.537152 4934 scope.go:117] "RemoveContainer" containerID="46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478" Oct 02 10:00:08 crc kubenswrapper[4934]: E1002 10:00:08.537557 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\": container with ID starting with 46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478 not found: ID does not exist" containerID="46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.537595 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478"} err="failed to get container status \"46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\": rpc error: code = NotFound desc = could not find container \"46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\": container with ID starting with 46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478 not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.537613 4934 scope.go:117] "RemoveContainer" containerID="354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.537963 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66"} err="failed to get container status \"354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66\": rpc error: code = NotFound desc = could not find container \"354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66\": container with ID starting with 354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66 not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.537984 4934 scope.go:117] "RemoveContainer" containerID="358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.538513 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe"} err="failed to get container status \"358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe\": rpc error: code = NotFound desc = could not find container \"358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe\": container with ID starting with 358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.538552 4934 scope.go:117] "RemoveContainer" containerID="2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.538987 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c"} err="failed to get container status \"2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\": rpc error: code = NotFound desc = could not find container \"2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\": container with ID starting with 2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.539009 4934 scope.go:117] "RemoveContainer" containerID="9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.539316 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea"} err="failed to get container status \"9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\": rpc error: code = NotFound desc = could not find container \"9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\": container with ID starting with 9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.539336 4934 scope.go:117] "RemoveContainer" containerID="c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.539562 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8"} err="failed to get container status \"c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\": rpc error: code = NotFound desc = could not find container \"c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\": container with ID starting with c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8 not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.539598 4934 scope.go:117] "RemoveContainer" containerID="389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.539836 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9"} err="failed to get container status \"389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\": rpc error: code = NotFound desc = could not find container \"389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\": container with ID starting with 389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9 not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.539855 4934 scope.go:117] "RemoveContainer" containerID="e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.540631 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b"} err="failed to get container status \"e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\": rpc error: code = NotFound desc = could not find container \"e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\": container with ID starting with e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.540683 4934 scope.go:117] "RemoveContainer" containerID="e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.541122 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019"} err="failed to get container status \"e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\": rpc error: code = NotFound desc = could not find container \"e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\": container with ID starting with e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019 not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.541142 4934 scope.go:117] "RemoveContainer" containerID="4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.542340 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb"} err="failed to get container status \"4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\": rpc error: code = NotFound desc = could not find container \"4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\": container with ID starting with 4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.542373 4934 scope.go:117] "RemoveContainer" containerID="46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.542742 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478"} err="failed to get container status \"46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\": rpc error: code = NotFound desc = could not find container \"46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\": container with ID starting with 46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478 not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.542768 4934 scope.go:117] "RemoveContainer" containerID="354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.543033 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66"} err="failed to get container status \"354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66\": rpc error: code = NotFound desc = could not find container \"354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66\": container with ID starting with 354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66 not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.543058 4934 scope.go:117] "RemoveContainer" containerID="358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.543394 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe"} err="failed to get container status \"358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe\": rpc error: code = NotFound desc = could not find container \"358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe\": container with ID starting with 358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.543411 4934 scope.go:117] "RemoveContainer" containerID="2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.543811 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c"} err="failed to get container status \"2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\": rpc error: code = NotFound desc = could not find container \"2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\": container with ID starting with 2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.543843 4934 scope.go:117] "RemoveContainer" containerID="9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.544186 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea"} err="failed to get container status \"9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\": rpc error: code = NotFound desc = could not find container \"9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\": container with ID starting with 9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.544203 4934 scope.go:117] "RemoveContainer" containerID="c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.544638 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8"} err="failed to get container status \"c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\": rpc error: code = NotFound desc = could not find container \"c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\": container with ID starting with c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8 not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.544660 4934 scope.go:117] "RemoveContainer" containerID="389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.544912 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9"} err="failed to get container status \"389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\": rpc error: code = NotFound desc = could not find container \"389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\": container with ID starting with 389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9 not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.544933 4934 scope.go:117] "RemoveContainer" containerID="e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.545356 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b"} err="failed to get container status \"e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\": rpc error: code = NotFound desc = could not find container \"e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\": container with ID starting with e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.545387 4934 scope.go:117] "RemoveContainer" containerID="e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.545801 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019"} err="failed to get container status \"e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\": rpc error: code = NotFound desc = could not find container \"e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\": container with ID starting with e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019 not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.545819 4934 scope.go:117] "RemoveContainer" containerID="4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.546063 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb"} err="failed to get container status \"4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\": rpc error: code = NotFound desc = could not find container \"4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\": container with ID starting with 4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.546085 4934 scope.go:117] "RemoveContainer" containerID="46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.546345 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478"} err="failed to get container status \"46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\": rpc error: code = NotFound desc = could not find container \"46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\": container with ID starting with 46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478 not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.546375 4934 scope.go:117] "RemoveContainer" containerID="354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.546700 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66"} err="failed to get container status \"354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66\": rpc error: code = NotFound desc = could not find container \"354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66\": container with ID starting with 354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66 not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.546729 4934 scope.go:117] "RemoveContainer" containerID="358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.546980 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe"} err="failed to get container status \"358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe\": rpc error: code = NotFound desc = could not find container \"358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe\": container with ID starting with 358d0d992919ef5ae02d4a472e089d875a8132ab4b2c2b63739293038c4d31fe not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.547003 4934 scope.go:117] "RemoveContainer" containerID="2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.547282 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c"} err="failed to get container status \"2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\": rpc error: code = NotFound desc = could not find container \"2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c\": container with ID starting with 2605a0486c035cc20b028f02ab4df52ca9374f67c79fce9c045d8ab83c46482c not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.547302 4934 scope.go:117] "RemoveContainer" containerID="9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.547543 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea"} err="failed to get container status \"9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\": rpc error: code = NotFound desc = could not find container \"9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea\": container with ID starting with 9460de32280a429e36eec881983dd6278eece3f2f67c61ac875a05ff703367ea not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.547567 4934 scope.go:117] "RemoveContainer" containerID="c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.547845 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8"} err="failed to get container status \"c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\": rpc error: code = NotFound desc = could not find container \"c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8\": container with ID starting with c55312df527ef6da7d42f0ffe8577038d0002633611213f60515f7fb1fdaf0c8 not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.547868 4934 scope.go:117] "RemoveContainer" containerID="389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.548139 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9"} err="failed to get container status \"389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\": rpc error: code = NotFound desc = could not find container \"389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9\": container with ID starting with 389092dc392e372bb0a581ed9bc22b55652beb755307f645383a69583c6370a9 not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.548159 4934 scope.go:117] "RemoveContainer" containerID="e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.548426 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b"} err="failed to get container status \"e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\": rpc error: code = NotFound desc = could not find container \"e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b\": container with ID starting with e1d99ff944b0e7ed6d0edb7fbbc68e941be4d8802702fe2b7e374391780e2e3b not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.548449 4934 scope.go:117] "RemoveContainer" containerID="e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.548732 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019"} err="failed to get container status \"e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\": rpc error: code = NotFound desc = could not find container \"e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019\": container with ID starting with e89e42fb2b52e1560b8b363636f54a5dfcfc26d468cf6895997224ca32c63019 not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.548752 4934 scope.go:117] "RemoveContainer" containerID="4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.549005 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb"} err="failed to get container status \"4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\": rpc error: code = NotFound desc = could not find container \"4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb\": container with ID starting with 4a894e5fdea86b1e03239d92e77a7097118b38759cf645a7505f8e7b8fc9b1bb not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.549029 4934 scope.go:117] "RemoveContainer" containerID="46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.549262 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478"} err="failed to get container status \"46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\": rpc error: code = NotFound desc = could not find container \"46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478\": container with ID starting with 46bf92e1eecf795dbb35cd47c94aadab3fc0618da57ae878e97aeec94ec5d478 not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.549279 4934 scope.go:117] "RemoveContainer" containerID="354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.549531 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66"} err="failed to get container status \"354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66\": rpc error: code = NotFound desc = could not find container \"354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66\": container with ID starting with 354d1718fdbd7fadc2a3e07c097c0b2443a920532f8290dbb5dc02dc78b50d66 not found: ID does not exist" Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.558252 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:08 crc kubenswrapper[4934]: W1002 10:00:08.583845 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4bffef6b_7a79_4c97_a874_686223e44047.slice/crio-3146ce46559b56b9c047c14bc5e50dc7e840b2b322f7e29c3d27867a5818b16e WatchSource:0}: Error finding container 3146ce46559b56b9c047c14bc5e50dc7e840b2b322f7e29c3d27867a5818b16e: Status 404 returned error can't find the container with id 3146ce46559b56b9c047c14bc5e50dc7e840b2b322f7e29c3d27867a5818b16e Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.654749 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-th4v6"] Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.658483 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-th4v6"] Oct 02 10:00:08 crc kubenswrapper[4934]: I1002 10:00:08.923695 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d4841b5-0469-461e-875c-25b9fe848141" path="/var/lib/kubelet/pods/1d4841b5-0469-461e-875c-25b9fe848141/volumes" Oct 02 10:00:09 crc kubenswrapper[4934]: I1002 10:00:09.332775 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-79fxg_73968247-67dd-48cc-88a1-64afac657412/kube-multus/1.log" Oct 02 10:00:09 crc kubenswrapper[4934]: I1002 10:00:09.333388 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-79fxg_73968247-67dd-48cc-88a1-64afac657412/kube-multus/0.log" Oct 02 10:00:09 crc kubenswrapper[4934]: I1002 10:00:09.333495 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-79fxg" event={"ID":"73968247-67dd-48cc-88a1-64afac657412","Type":"ContainerStarted","Data":"81bbec58862a692829142b029a25d89c551d380c2fba2888e6af2c37e9ab835a"} Oct 02 10:00:09 crc kubenswrapper[4934]: I1002 10:00:09.336348 4934 generic.go:334] "Generic (PLEG): container finished" podID="4bffef6b-7a79-4c97-a874-686223e44047" containerID="c3226f5c2e4e15837927e2ff7dbcd711127da8ea5f1dfd08e7e954cf645e4904" exitCode=0 Oct 02 10:00:09 crc kubenswrapper[4934]: I1002 10:00:09.336443 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" event={"ID":"4bffef6b-7a79-4c97-a874-686223e44047","Type":"ContainerDied","Data":"c3226f5c2e4e15837927e2ff7dbcd711127da8ea5f1dfd08e7e954cf645e4904"} Oct 02 10:00:09 crc kubenswrapper[4934]: I1002 10:00:09.336490 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" event={"ID":"4bffef6b-7a79-4c97-a874-686223e44047","Type":"ContainerStarted","Data":"3146ce46559b56b9c047c14bc5e50dc7e840b2b322f7e29c3d27867a5818b16e"} Oct 02 10:00:09 crc kubenswrapper[4934]: I1002 10:00:09.457121 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl" Oct 02 10:00:09 crc kubenswrapper[4934]: I1002 10:00:09.464504 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rlnr\" (UniqueName: \"kubernetes.io/projected/89aa6f88-b397-4574-8ad9-6c855a2c8511-kube-api-access-9rlnr\") pod \"89aa6f88-b397-4574-8ad9-6c855a2c8511\" (UID: \"89aa6f88-b397-4574-8ad9-6c855a2c8511\") " Oct 02 10:00:09 crc kubenswrapper[4934]: I1002 10:00:09.464594 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/89aa6f88-b397-4574-8ad9-6c855a2c8511-bundle\") pod \"89aa6f88-b397-4574-8ad9-6c855a2c8511\" (UID: \"89aa6f88-b397-4574-8ad9-6c855a2c8511\") " Oct 02 10:00:09 crc kubenswrapper[4934]: I1002 10:00:09.464637 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/89aa6f88-b397-4574-8ad9-6c855a2c8511-util\") pod \"89aa6f88-b397-4574-8ad9-6c855a2c8511\" (UID: \"89aa6f88-b397-4574-8ad9-6c855a2c8511\") " Oct 02 10:00:09 crc kubenswrapper[4934]: I1002 10:00:09.465496 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89aa6f88-b397-4574-8ad9-6c855a2c8511-bundle" (OuterVolumeSpecName: "bundle") pod "89aa6f88-b397-4574-8ad9-6c855a2c8511" (UID: "89aa6f88-b397-4574-8ad9-6c855a2c8511"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:00:09 crc kubenswrapper[4934]: I1002 10:00:09.468941 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89aa6f88-b397-4574-8ad9-6c855a2c8511-kube-api-access-9rlnr" (OuterVolumeSpecName: "kube-api-access-9rlnr") pod "89aa6f88-b397-4574-8ad9-6c855a2c8511" (UID: "89aa6f88-b397-4574-8ad9-6c855a2c8511"). InnerVolumeSpecName "kube-api-access-9rlnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:00:09 crc kubenswrapper[4934]: I1002 10:00:09.566948 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rlnr\" (UniqueName: \"kubernetes.io/projected/89aa6f88-b397-4574-8ad9-6c855a2c8511-kube-api-access-9rlnr\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:09 crc kubenswrapper[4934]: I1002 10:00:09.566998 4934 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/89aa6f88-b397-4574-8ad9-6c855a2c8511-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:09 crc kubenswrapper[4934]: I1002 10:00:09.771053 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89aa6f88-b397-4574-8ad9-6c855a2c8511-util" (OuterVolumeSpecName: "util") pod "89aa6f88-b397-4574-8ad9-6c855a2c8511" (UID: "89aa6f88-b397-4574-8ad9-6c855a2c8511"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:00:09 crc kubenswrapper[4934]: I1002 10:00:09.869823 4934 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/89aa6f88-b397-4574-8ad9-6c855a2c8511-util\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:10 crc kubenswrapper[4934]: I1002 10:00:10.353609 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" event={"ID":"4bffef6b-7a79-4c97-a874-686223e44047","Type":"ContainerStarted","Data":"875275eb6ac54c9d746b16941291d3caf0498cbb0a32303cdfcadc61edc8a123"} Oct 02 10:00:10 crc kubenswrapper[4934]: I1002 10:00:10.353657 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" event={"ID":"4bffef6b-7a79-4c97-a874-686223e44047","Type":"ContainerStarted","Data":"e49a9baa19dc450393ca5a30154c8ba0a580ff099cc06cb3a4a7d6bcf33c9d5f"} Oct 02 10:00:10 crc kubenswrapper[4934]: I1002 10:00:10.353670 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" event={"ID":"4bffef6b-7a79-4c97-a874-686223e44047","Type":"ContainerStarted","Data":"c158389c615d9f01a08881c76b5daec6b4041095024ccae9cdb5fbfe5a5ec910"} Oct 02 10:00:10 crc kubenswrapper[4934]: I1002 10:00:10.353682 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" event={"ID":"4bffef6b-7a79-4c97-a874-686223e44047","Type":"ContainerStarted","Data":"9cd1c56afd6a25f14aaabb7363d0214c1b9c46d74323351ca7c5e6a733b5cae9"} Oct 02 10:00:10 crc kubenswrapper[4934]: I1002 10:00:10.353695 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" event={"ID":"4bffef6b-7a79-4c97-a874-686223e44047","Type":"ContainerStarted","Data":"d7d97631e23e8cf73d86c4c072ff60e187d0d708195903207aa41568a5cd3c72"} Oct 02 10:00:10 crc kubenswrapper[4934]: I1002 10:00:10.356810 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl" event={"ID":"89aa6f88-b397-4574-8ad9-6c855a2c8511","Type":"ContainerDied","Data":"b7a7213977b6eb6117a3c80ce003f759674f578dc92e582fca1435e58eae077a"} Oct 02 10:00:10 crc kubenswrapper[4934]: I1002 10:00:10.356833 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7a7213977b6eb6117a3c80ce003f759674f578dc92e582fca1435e58eae077a" Oct 02 10:00:10 crc kubenswrapper[4934]: I1002 10:00:10.356966 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl" Oct 02 10:00:11 crc kubenswrapper[4934]: I1002 10:00:11.368902 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" event={"ID":"4bffef6b-7a79-4c97-a874-686223e44047","Type":"ContainerStarted","Data":"6a01b645797895df2d214c009275c274b0f052ff10b9d88b0776f5d63a2297b8"} Oct 02 10:00:13 crc kubenswrapper[4934]: I1002 10:00:13.383161 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" event={"ID":"4bffef6b-7a79-4c97-a874-686223e44047","Type":"ContainerStarted","Data":"7fb8956aeeac3df525251fa3f4c3aca7f0e15abebeff921c2936ff69299e7d86"} Oct 02 10:00:14 crc kubenswrapper[4934]: I1002 10:00:14.304850 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-xcdzm"] Oct 02 10:00:14 crc kubenswrapper[4934]: E1002 10:00:14.305070 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89aa6f88-b397-4574-8ad9-6c855a2c8511" containerName="extract" Oct 02 10:00:14 crc kubenswrapper[4934]: I1002 10:00:14.305084 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="89aa6f88-b397-4574-8ad9-6c855a2c8511" containerName="extract" Oct 02 10:00:14 crc kubenswrapper[4934]: E1002 10:00:14.305103 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89aa6f88-b397-4574-8ad9-6c855a2c8511" containerName="pull" Oct 02 10:00:14 crc kubenswrapper[4934]: I1002 10:00:14.305110 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="89aa6f88-b397-4574-8ad9-6c855a2c8511" containerName="pull" Oct 02 10:00:14 crc kubenswrapper[4934]: E1002 10:00:14.305124 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89aa6f88-b397-4574-8ad9-6c855a2c8511" containerName="util" Oct 02 10:00:14 crc kubenswrapper[4934]: I1002 10:00:14.305132 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="89aa6f88-b397-4574-8ad9-6c855a2c8511" containerName="util" Oct 02 10:00:14 crc kubenswrapper[4934]: I1002 10:00:14.305256 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="89aa6f88-b397-4574-8ad9-6c855a2c8511" containerName="extract" Oct 02 10:00:14 crc kubenswrapper[4934]: I1002 10:00:14.305705 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-xcdzm" Oct 02 10:00:14 crc kubenswrapper[4934]: I1002 10:00:14.307817 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-pk9ss" Oct 02 10:00:14 crc kubenswrapper[4934]: I1002 10:00:14.308159 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 02 10:00:14 crc kubenswrapper[4934]: I1002 10:00:14.309365 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 02 10:00:14 crc kubenswrapper[4934]: I1002 10:00:14.325074 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bz8pd\" (UniqueName: \"kubernetes.io/projected/b9aa6395-2651-499c-afe7-ec95844b75d3-kube-api-access-bz8pd\") pod \"nmstate-operator-858ddd8f98-xcdzm\" (UID: \"b9aa6395-2651-499c-afe7-ec95844b75d3\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-xcdzm" Oct 02 10:00:14 crc kubenswrapper[4934]: I1002 10:00:14.426521 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bz8pd\" (UniqueName: \"kubernetes.io/projected/b9aa6395-2651-499c-afe7-ec95844b75d3-kube-api-access-bz8pd\") pod \"nmstate-operator-858ddd8f98-xcdzm\" (UID: \"b9aa6395-2651-499c-afe7-ec95844b75d3\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-xcdzm" Oct 02 10:00:14 crc kubenswrapper[4934]: I1002 10:00:14.447938 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bz8pd\" (UniqueName: \"kubernetes.io/projected/b9aa6395-2651-499c-afe7-ec95844b75d3-kube-api-access-bz8pd\") pod \"nmstate-operator-858ddd8f98-xcdzm\" (UID: \"b9aa6395-2651-499c-afe7-ec95844b75d3\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-xcdzm" Oct 02 10:00:14 crc kubenswrapper[4934]: I1002 10:00:14.630760 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-xcdzm" Oct 02 10:00:14 crc kubenswrapper[4934]: E1002 10:00:14.656093 4934 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_nmstate-operator-858ddd8f98-xcdzm_openshift-nmstate_b9aa6395-2651-499c-afe7-ec95844b75d3_0(da757d85916b3cdb967c53cca882cbb95bf34c272c6e102e78528a925e58ffbe): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 10:00:14 crc kubenswrapper[4934]: E1002 10:00:14.656525 4934 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_nmstate-operator-858ddd8f98-xcdzm_openshift-nmstate_b9aa6395-2651-499c-afe7-ec95844b75d3_0(da757d85916b3cdb967c53cca882cbb95bf34c272c6e102e78528a925e58ffbe): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-nmstate/nmstate-operator-858ddd8f98-xcdzm" Oct 02 10:00:14 crc kubenswrapper[4934]: E1002 10:00:14.656718 4934 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_nmstate-operator-858ddd8f98-xcdzm_openshift-nmstate_b9aa6395-2651-499c-afe7-ec95844b75d3_0(da757d85916b3cdb967c53cca882cbb95bf34c272c6e102e78528a925e58ffbe): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-nmstate/nmstate-operator-858ddd8f98-xcdzm" Oct 02 10:00:14 crc kubenswrapper[4934]: E1002 10:00:14.656873 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"nmstate-operator-858ddd8f98-xcdzm_openshift-nmstate(b9aa6395-2651-499c-afe7-ec95844b75d3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"nmstate-operator-858ddd8f98-xcdzm_openshift-nmstate(b9aa6395-2651-499c-afe7-ec95844b75d3)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_nmstate-operator-858ddd8f98-xcdzm_openshift-nmstate_b9aa6395-2651-499c-afe7-ec95844b75d3_0(da757d85916b3cdb967c53cca882cbb95bf34c272c6e102e78528a925e58ffbe): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-nmstate/nmstate-operator-858ddd8f98-xcdzm" podUID="b9aa6395-2651-499c-afe7-ec95844b75d3" Oct 02 10:00:15 crc kubenswrapper[4934]: I1002 10:00:15.411503 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" event={"ID":"4bffef6b-7a79-4c97-a874-686223e44047","Type":"ContainerStarted","Data":"16ae7e471b7811a9ec87a9680983f875ca4deffba8657d822199c451c0d1fb4c"} Oct 02 10:00:15 crc kubenswrapper[4934]: I1002 10:00:15.411947 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:15 crc kubenswrapper[4934]: I1002 10:00:15.411997 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:15 crc kubenswrapper[4934]: I1002 10:00:15.412011 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:15 crc kubenswrapper[4934]: I1002 10:00:15.442954 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" podStartSLOduration=7.442936508 podStartE2EDuration="7.442936508s" podCreationTimestamp="2025-10-02 10:00:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:00:15.442630589 +0000 UTC m=+687.195272131" watchObservedRunningTime="2025-10-02 10:00:15.442936508 +0000 UTC m=+687.195578030" Oct 02 10:00:15 crc kubenswrapper[4934]: I1002 10:00:15.447705 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:15 crc kubenswrapper[4934]: I1002 10:00:15.454490 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:15 crc kubenswrapper[4934]: I1002 10:00:15.722611 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-xcdzm"] Oct 02 10:00:15 crc kubenswrapper[4934]: I1002 10:00:15.722721 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-xcdzm" Oct 02 10:00:15 crc kubenswrapper[4934]: I1002 10:00:15.723179 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-xcdzm" Oct 02 10:00:15 crc kubenswrapper[4934]: E1002 10:00:15.748599 4934 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_nmstate-operator-858ddd8f98-xcdzm_openshift-nmstate_b9aa6395-2651-499c-afe7-ec95844b75d3_0(c889560a112ccdbe23c7e03fe28f180bb1087d41c62b030801cc1b0c009566ef): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 10:00:15 crc kubenswrapper[4934]: E1002 10:00:15.748670 4934 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_nmstate-operator-858ddd8f98-xcdzm_openshift-nmstate_b9aa6395-2651-499c-afe7-ec95844b75d3_0(c889560a112ccdbe23c7e03fe28f180bb1087d41c62b030801cc1b0c009566ef): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-nmstate/nmstate-operator-858ddd8f98-xcdzm" Oct 02 10:00:15 crc kubenswrapper[4934]: E1002 10:00:15.748691 4934 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_nmstate-operator-858ddd8f98-xcdzm_openshift-nmstate_b9aa6395-2651-499c-afe7-ec95844b75d3_0(c889560a112ccdbe23c7e03fe28f180bb1087d41c62b030801cc1b0c009566ef): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-nmstate/nmstate-operator-858ddd8f98-xcdzm" Oct 02 10:00:15 crc kubenswrapper[4934]: E1002 10:00:15.748737 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"nmstate-operator-858ddd8f98-xcdzm_openshift-nmstate(b9aa6395-2651-499c-afe7-ec95844b75d3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"nmstate-operator-858ddd8f98-xcdzm_openshift-nmstate(b9aa6395-2651-499c-afe7-ec95844b75d3)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_nmstate-operator-858ddd8f98-xcdzm_openshift-nmstate_b9aa6395-2651-499c-afe7-ec95844b75d3_0(c889560a112ccdbe23c7e03fe28f180bb1087d41c62b030801cc1b0c009566ef): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-nmstate/nmstate-operator-858ddd8f98-xcdzm" podUID="b9aa6395-2651-499c-afe7-ec95844b75d3" Oct 02 10:00:26 crc kubenswrapper[4934]: I1002 10:00:26.913020 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-xcdzm" Oct 02 10:00:26 crc kubenswrapper[4934]: I1002 10:00:26.914358 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-xcdzm" Oct 02 10:00:27 crc kubenswrapper[4934]: I1002 10:00:27.131805 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-xcdzm"] Oct 02 10:00:27 crc kubenswrapper[4934]: W1002 10:00:27.134994 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9aa6395_2651_499c_afe7_ec95844b75d3.slice/crio-89c7a514d0feb92632525128611d71baa7ad2db5091bf7578cea3c6aa0d0cf66 WatchSource:0}: Error finding container 89c7a514d0feb92632525128611d71baa7ad2db5091bf7578cea3c6aa0d0cf66: Status 404 returned error can't find the container with id 89c7a514d0feb92632525128611d71baa7ad2db5091bf7578cea3c6aa0d0cf66 Oct 02 10:00:27 crc kubenswrapper[4934]: I1002 10:00:27.476278 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-xcdzm" event={"ID":"b9aa6395-2651-499c-afe7-ec95844b75d3","Type":"ContainerStarted","Data":"89c7a514d0feb92632525128611d71baa7ad2db5091bf7578cea3c6aa0d0cf66"} Oct 02 10:00:30 crc kubenswrapper[4934]: I1002 10:00:30.499997 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-xcdzm" event={"ID":"b9aa6395-2651-499c-afe7-ec95844b75d3","Type":"ContainerStarted","Data":"f0e0cd0776313740ed8759ad052af627b252031dc7e4cf9b6914ea5c334731da"} Oct 02 10:00:30 crc kubenswrapper[4934]: I1002 10:00:30.522663 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-xcdzm" podStartSLOduration=14.28819457 podStartE2EDuration="16.522636404s" podCreationTimestamp="2025-10-02 10:00:14 +0000 UTC" firstStartedPulling="2025-10-02 10:00:27.137525649 +0000 UTC m=+698.890167171" lastFinishedPulling="2025-10-02 10:00:29.371967483 +0000 UTC m=+701.124609005" observedRunningTime="2025-10-02 10:00:30.520522525 +0000 UTC m=+702.273164047" watchObservedRunningTime="2025-10-02 10:00:30.522636404 +0000 UTC m=+702.275277956" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.514017 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-lwrzm"] Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.515044 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-lwrzm" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.517260 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-wfjnd" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.557194 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-w6z5d"] Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.558374 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-w6z5d" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.560466 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.566145 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lp8k4\" (UniqueName: \"kubernetes.io/projected/d705acd3-4167-430e-b33d-5966e485a9ce-kube-api-access-lp8k4\") pod \"nmstate-metrics-fdff9cb8d-lwrzm\" (UID: \"d705acd3-4167-430e-b33d-5966e485a9ce\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-lwrzm" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.583716 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-w6z5d"] Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.598804 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-lwrzm"] Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.601300 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-mb8mr"] Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.601921 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-mb8mr" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.667524 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/687810e0-5141-42c9-a40b-5f04d045bd5d-dbus-socket\") pod \"nmstate-handler-mb8mr\" (UID: \"687810e0-5141-42c9-a40b-5f04d045bd5d\") " pod="openshift-nmstate/nmstate-handler-mb8mr" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.667795 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jp26s\" (UniqueName: \"kubernetes.io/projected/687810e0-5141-42c9-a40b-5f04d045bd5d-kube-api-access-jp26s\") pod \"nmstate-handler-mb8mr\" (UID: \"687810e0-5141-42c9-a40b-5f04d045bd5d\") " pod="openshift-nmstate/nmstate-handler-mb8mr" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.667927 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/687810e0-5141-42c9-a40b-5f04d045bd5d-nmstate-lock\") pod \"nmstate-handler-mb8mr\" (UID: \"687810e0-5141-42c9-a40b-5f04d045bd5d\") " pod="openshift-nmstate/nmstate-handler-mb8mr" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.667995 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpxv8\" (UniqueName: \"kubernetes.io/projected/f45d40d5-581b-47e0-a25e-9c550f92f689-kube-api-access-rpxv8\") pod \"nmstate-webhook-6cdbc54649-w6z5d\" (UID: \"f45d40d5-581b-47e0-a25e-9c550f92f689\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-w6z5d" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.668074 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lp8k4\" (UniqueName: \"kubernetes.io/projected/d705acd3-4167-430e-b33d-5966e485a9ce-kube-api-access-lp8k4\") pod \"nmstate-metrics-fdff9cb8d-lwrzm\" (UID: \"d705acd3-4167-430e-b33d-5966e485a9ce\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-lwrzm" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.668149 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f45d40d5-581b-47e0-a25e-9c550f92f689-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-w6z5d\" (UID: \"f45d40d5-581b-47e0-a25e-9c550f92f689\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-w6z5d" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.668248 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/687810e0-5141-42c9-a40b-5f04d045bd5d-ovs-socket\") pod \"nmstate-handler-mb8mr\" (UID: \"687810e0-5141-42c9-a40b-5f04d045bd5d\") " pod="openshift-nmstate/nmstate-handler-mb8mr" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.706926 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lp8k4\" (UniqueName: \"kubernetes.io/projected/d705acd3-4167-430e-b33d-5966e485a9ce-kube-api-access-lp8k4\") pod \"nmstate-metrics-fdff9cb8d-lwrzm\" (UID: \"d705acd3-4167-430e-b33d-5966e485a9ce\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-lwrzm" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.714289 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-msxrz"] Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.716142 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-msxrz" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.717937 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.717965 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.719608 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-8fbjj" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.728162 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-msxrz"] Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.770108 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ef8e1eb9-668c-4542-9551-8a02b13222b9-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-msxrz\" (UID: \"ef8e1eb9-668c-4542-9551-8a02b13222b9\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-msxrz" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.770487 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jp26s\" (UniqueName: \"kubernetes.io/projected/687810e0-5141-42c9-a40b-5f04d045bd5d-kube-api-access-jp26s\") pod \"nmstate-handler-mb8mr\" (UID: \"687810e0-5141-42c9-a40b-5f04d045bd5d\") " pod="openshift-nmstate/nmstate-handler-mb8mr" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.770613 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/687810e0-5141-42c9-a40b-5f04d045bd5d-nmstate-lock\") pod \"nmstate-handler-mb8mr\" (UID: \"687810e0-5141-42c9-a40b-5f04d045bd5d\") " pod="openshift-nmstate/nmstate-handler-mb8mr" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.770696 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/687810e0-5141-42c9-a40b-5f04d045bd5d-nmstate-lock\") pod \"nmstate-handler-mb8mr\" (UID: \"687810e0-5141-42c9-a40b-5f04d045bd5d\") " pod="openshift-nmstate/nmstate-handler-mb8mr" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.771177 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpxv8\" (UniqueName: \"kubernetes.io/projected/f45d40d5-581b-47e0-a25e-9c550f92f689-kube-api-access-rpxv8\") pod \"nmstate-webhook-6cdbc54649-w6z5d\" (UID: \"f45d40d5-581b-47e0-a25e-9c550f92f689\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-w6z5d" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.771318 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f45d40d5-581b-47e0-a25e-9c550f92f689-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-w6z5d\" (UID: \"f45d40d5-581b-47e0-a25e-9c550f92f689\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-w6z5d" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.771490 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/687810e0-5141-42c9-a40b-5f04d045bd5d-ovs-socket\") pod \"nmstate-handler-mb8mr\" (UID: \"687810e0-5141-42c9-a40b-5f04d045bd5d\") " pod="openshift-nmstate/nmstate-handler-mb8mr" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.771671 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxwx4\" (UniqueName: \"kubernetes.io/projected/ef8e1eb9-668c-4542-9551-8a02b13222b9-kube-api-access-xxwx4\") pod \"nmstate-console-plugin-6b874cbd85-msxrz\" (UID: \"ef8e1eb9-668c-4542-9551-8a02b13222b9\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-msxrz" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.771614 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/687810e0-5141-42c9-a40b-5f04d045bd5d-ovs-socket\") pod \"nmstate-handler-mb8mr\" (UID: \"687810e0-5141-42c9-a40b-5f04d045bd5d\") " pod="openshift-nmstate/nmstate-handler-mb8mr" Oct 02 10:00:31 crc kubenswrapper[4934]: E1002 10:00:31.771450 4934 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 02 10:00:31 crc kubenswrapper[4934]: E1002 10:00:31.771989 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f45d40d5-581b-47e0-a25e-9c550f92f689-tls-key-pair podName:f45d40d5-581b-47e0-a25e-9c550f92f689 nodeName:}" failed. No retries permitted until 2025-10-02 10:00:32.271972112 +0000 UTC m=+704.024613624 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/f45d40d5-581b-47e0-a25e-9c550f92f689-tls-key-pair") pod "nmstate-webhook-6cdbc54649-w6z5d" (UID: "f45d40d5-581b-47e0-a25e-9c550f92f689") : secret "openshift-nmstate-webhook" not found Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.771813 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/687810e0-5141-42c9-a40b-5f04d045bd5d-dbus-socket\") pod \"nmstate-handler-mb8mr\" (UID: \"687810e0-5141-42c9-a40b-5f04d045bd5d\") " pod="openshift-nmstate/nmstate-handler-mb8mr" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.772168 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ef8e1eb9-668c-4542-9551-8a02b13222b9-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-msxrz\" (UID: \"ef8e1eb9-668c-4542-9551-8a02b13222b9\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-msxrz" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.772229 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/687810e0-5141-42c9-a40b-5f04d045bd5d-dbus-socket\") pod \"nmstate-handler-mb8mr\" (UID: \"687810e0-5141-42c9-a40b-5f04d045bd5d\") " pod="openshift-nmstate/nmstate-handler-mb8mr" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.785872 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jp26s\" (UniqueName: \"kubernetes.io/projected/687810e0-5141-42c9-a40b-5f04d045bd5d-kube-api-access-jp26s\") pod \"nmstate-handler-mb8mr\" (UID: \"687810e0-5141-42c9-a40b-5f04d045bd5d\") " pod="openshift-nmstate/nmstate-handler-mb8mr" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.792412 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpxv8\" (UniqueName: \"kubernetes.io/projected/f45d40d5-581b-47e0-a25e-9c550f92f689-kube-api-access-rpxv8\") pod \"nmstate-webhook-6cdbc54649-w6z5d\" (UID: \"f45d40d5-581b-47e0-a25e-9c550f92f689\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-w6z5d" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.834315 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-lwrzm" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.873187 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ef8e1eb9-668c-4542-9551-8a02b13222b9-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-msxrz\" (UID: \"ef8e1eb9-668c-4542-9551-8a02b13222b9\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-msxrz" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.873248 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ef8e1eb9-668c-4542-9551-8a02b13222b9-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-msxrz\" (UID: \"ef8e1eb9-668c-4542-9551-8a02b13222b9\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-msxrz" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.873355 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxwx4\" (UniqueName: \"kubernetes.io/projected/ef8e1eb9-668c-4542-9551-8a02b13222b9-kube-api-access-xxwx4\") pod \"nmstate-console-plugin-6b874cbd85-msxrz\" (UID: \"ef8e1eb9-668c-4542-9551-8a02b13222b9\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-msxrz" Oct 02 10:00:31 crc kubenswrapper[4934]: E1002 10:00:31.873702 4934 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Oct 02 10:00:31 crc kubenswrapper[4934]: E1002 10:00:31.873855 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ef8e1eb9-668c-4542-9551-8a02b13222b9-plugin-serving-cert podName:ef8e1eb9-668c-4542-9551-8a02b13222b9 nodeName:}" failed. No retries permitted until 2025-10-02 10:00:32.373829758 +0000 UTC m=+704.126471290 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/ef8e1eb9-668c-4542-9551-8a02b13222b9-plugin-serving-cert") pod "nmstate-console-plugin-6b874cbd85-msxrz" (UID: "ef8e1eb9-668c-4542-9551-8a02b13222b9") : secret "plugin-serving-cert" not found Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.874443 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ef8e1eb9-668c-4542-9551-8a02b13222b9-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-msxrz\" (UID: \"ef8e1eb9-668c-4542-9551-8a02b13222b9\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-msxrz" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.895313 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxwx4\" (UniqueName: \"kubernetes.io/projected/ef8e1eb9-668c-4542-9551-8a02b13222b9-kube-api-access-xxwx4\") pod \"nmstate-console-plugin-6b874cbd85-msxrz\" (UID: \"ef8e1eb9-668c-4542-9551-8a02b13222b9\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-msxrz" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.916291 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-mb8mr" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.929792 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7cc8bd7b4-jmbrc"] Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.930416 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7cc8bd7b4-jmbrc" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.946885 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7cc8bd7b4-jmbrc"] Oct 02 10:00:31 crc kubenswrapper[4934]: W1002 10:00:31.948299 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod687810e0_5141_42c9_a40b_5f04d045bd5d.slice/crio-689865ed5b92193654e0e1b62d329c8f763e773776319634149eef73aa3b6cbb WatchSource:0}: Error finding container 689865ed5b92193654e0e1b62d329c8f763e773776319634149eef73aa3b6cbb: Status 404 returned error can't find the container with id 689865ed5b92193654e0e1b62d329c8f763e773776319634149eef73aa3b6cbb Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.974820 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c1b9dffc-787b-4750-8242-e2716633ca40-trusted-ca-bundle\") pod \"console-7cc8bd7b4-jmbrc\" (UID: \"c1b9dffc-787b-4750-8242-e2716633ca40\") " pod="openshift-console/console-7cc8bd7b4-jmbrc" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.975099 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c1b9dffc-787b-4750-8242-e2716633ca40-service-ca\") pod \"console-7cc8bd7b4-jmbrc\" (UID: \"c1b9dffc-787b-4750-8242-e2716633ca40\") " pod="openshift-console/console-7cc8bd7b4-jmbrc" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.975125 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c1b9dffc-787b-4750-8242-e2716633ca40-console-oauth-config\") pod \"console-7cc8bd7b4-jmbrc\" (UID: \"c1b9dffc-787b-4750-8242-e2716633ca40\") " pod="openshift-console/console-7cc8bd7b4-jmbrc" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.975195 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c1b9dffc-787b-4750-8242-e2716633ca40-oauth-serving-cert\") pod \"console-7cc8bd7b4-jmbrc\" (UID: \"c1b9dffc-787b-4750-8242-e2716633ca40\") " pod="openshift-console/console-7cc8bd7b4-jmbrc" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.975220 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vc27k\" (UniqueName: \"kubernetes.io/projected/c1b9dffc-787b-4750-8242-e2716633ca40-kube-api-access-vc27k\") pod \"console-7cc8bd7b4-jmbrc\" (UID: \"c1b9dffc-787b-4750-8242-e2716633ca40\") " pod="openshift-console/console-7cc8bd7b4-jmbrc" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.975248 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c1b9dffc-787b-4750-8242-e2716633ca40-console-config\") pod \"console-7cc8bd7b4-jmbrc\" (UID: \"c1b9dffc-787b-4750-8242-e2716633ca40\") " pod="openshift-console/console-7cc8bd7b4-jmbrc" Oct 02 10:00:31 crc kubenswrapper[4934]: I1002 10:00:31.975341 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c1b9dffc-787b-4750-8242-e2716633ca40-console-serving-cert\") pod \"console-7cc8bd7b4-jmbrc\" (UID: \"c1b9dffc-787b-4750-8242-e2716633ca40\") " pod="openshift-console/console-7cc8bd7b4-jmbrc" Oct 02 10:00:32 crc kubenswrapper[4934]: I1002 10:00:32.076080 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c1b9dffc-787b-4750-8242-e2716633ca40-trusted-ca-bundle\") pod \"console-7cc8bd7b4-jmbrc\" (UID: \"c1b9dffc-787b-4750-8242-e2716633ca40\") " pod="openshift-console/console-7cc8bd7b4-jmbrc" Oct 02 10:00:32 crc kubenswrapper[4934]: I1002 10:00:32.076153 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c1b9dffc-787b-4750-8242-e2716633ca40-console-oauth-config\") pod \"console-7cc8bd7b4-jmbrc\" (UID: \"c1b9dffc-787b-4750-8242-e2716633ca40\") " pod="openshift-console/console-7cc8bd7b4-jmbrc" Oct 02 10:00:32 crc kubenswrapper[4934]: I1002 10:00:32.076172 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c1b9dffc-787b-4750-8242-e2716633ca40-service-ca\") pod \"console-7cc8bd7b4-jmbrc\" (UID: \"c1b9dffc-787b-4750-8242-e2716633ca40\") " pod="openshift-console/console-7cc8bd7b4-jmbrc" Oct 02 10:00:32 crc kubenswrapper[4934]: I1002 10:00:32.076233 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c1b9dffc-787b-4750-8242-e2716633ca40-oauth-serving-cert\") pod \"console-7cc8bd7b4-jmbrc\" (UID: \"c1b9dffc-787b-4750-8242-e2716633ca40\") " pod="openshift-console/console-7cc8bd7b4-jmbrc" Oct 02 10:00:32 crc kubenswrapper[4934]: I1002 10:00:32.076252 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vc27k\" (UniqueName: \"kubernetes.io/projected/c1b9dffc-787b-4750-8242-e2716633ca40-kube-api-access-vc27k\") pod \"console-7cc8bd7b4-jmbrc\" (UID: \"c1b9dffc-787b-4750-8242-e2716633ca40\") " pod="openshift-console/console-7cc8bd7b4-jmbrc" Oct 02 10:00:32 crc kubenswrapper[4934]: I1002 10:00:32.076278 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c1b9dffc-787b-4750-8242-e2716633ca40-console-config\") pod \"console-7cc8bd7b4-jmbrc\" (UID: \"c1b9dffc-787b-4750-8242-e2716633ca40\") " pod="openshift-console/console-7cc8bd7b4-jmbrc" Oct 02 10:00:32 crc kubenswrapper[4934]: I1002 10:00:32.076329 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c1b9dffc-787b-4750-8242-e2716633ca40-console-serving-cert\") pod \"console-7cc8bd7b4-jmbrc\" (UID: \"c1b9dffc-787b-4750-8242-e2716633ca40\") " pod="openshift-console/console-7cc8bd7b4-jmbrc" Oct 02 10:00:32 crc kubenswrapper[4934]: I1002 10:00:32.077899 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c1b9dffc-787b-4750-8242-e2716633ca40-trusted-ca-bundle\") pod \"console-7cc8bd7b4-jmbrc\" (UID: \"c1b9dffc-787b-4750-8242-e2716633ca40\") " pod="openshift-console/console-7cc8bd7b4-jmbrc" Oct 02 10:00:32 crc kubenswrapper[4934]: I1002 10:00:32.078232 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c1b9dffc-787b-4750-8242-e2716633ca40-oauth-serving-cert\") pod \"console-7cc8bd7b4-jmbrc\" (UID: \"c1b9dffc-787b-4750-8242-e2716633ca40\") " pod="openshift-console/console-7cc8bd7b4-jmbrc" Oct 02 10:00:32 crc kubenswrapper[4934]: I1002 10:00:32.078412 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c1b9dffc-787b-4750-8242-e2716633ca40-service-ca\") pod \"console-7cc8bd7b4-jmbrc\" (UID: \"c1b9dffc-787b-4750-8242-e2716633ca40\") " pod="openshift-console/console-7cc8bd7b4-jmbrc" Oct 02 10:00:32 crc kubenswrapper[4934]: I1002 10:00:32.078443 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c1b9dffc-787b-4750-8242-e2716633ca40-console-config\") pod \"console-7cc8bd7b4-jmbrc\" (UID: \"c1b9dffc-787b-4750-8242-e2716633ca40\") " pod="openshift-console/console-7cc8bd7b4-jmbrc" Oct 02 10:00:32 crc kubenswrapper[4934]: I1002 10:00:32.080693 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c1b9dffc-787b-4750-8242-e2716633ca40-console-serving-cert\") pod \"console-7cc8bd7b4-jmbrc\" (UID: \"c1b9dffc-787b-4750-8242-e2716633ca40\") " pod="openshift-console/console-7cc8bd7b4-jmbrc" Oct 02 10:00:32 crc kubenswrapper[4934]: I1002 10:00:32.080935 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c1b9dffc-787b-4750-8242-e2716633ca40-console-oauth-config\") pod \"console-7cc8bd7b4-jmbrc\" (UID: \"c1b9dffc-787b-4750-8242-e2716633ca40\") " pod="openshift-console/console-7cc8bd7b4-jmbrc" Oct 02 10:00:32 crc kubenswrapper[4934]: I1002 10:00:32.093952 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-lwrzm"] Oct 02 10:00:32 crc kubenswrapper[4934]: I1002 10:00:32.102889 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vc27k\" (UniqueName: \"kubernetes.io/projected/c1b9dffc-787b-4750-8242-e2716633ca40-kube-api-access-vc27k\") pod \"console-7cc8bd7b4-jmbrc\" (UID: \"c1b9dffc-787b-4750-8242-e2716633ca40\") " pod="openshift-console/console-7cc8bd7b4-jmbrc" Oct 02 10:00:32 crc kubenswrapper[4934]: I1002 10:00:32.246724 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7cc8bd7b4-jmbrc" Oct 02 10:00:32 crc kubenswrapper[4934]: I1002 10:00:32.279079 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f45d40d5-581b-47e0-a25e-9c550f92f689-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-w6z5d\" (UID: \"f45d40d5-581b-47e0-a25e-9c550f92f689\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-w6z5d" Oct 02 10:00:32 crc kubenswrapper[4934]: I1002 10:00:32.285691 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f45d40d5-581b-47e0-a25e-9c550f92f689-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-w6z5d\" (UID: \"f45d40d5-581b-47e0-a25e-9c550f92f689\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-w6z5d" Oct 02 10:00:32 crc kubenswrapper[4934]: I1002 10:00:32.380851 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ef8e1eb9-668c-4542-9551-8a02b13222b9-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-msxrz\" (UID: \"ef8e1eb9-668c-4542-9551-8a02b13222b9\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-msxrz" Oct 02 10:00:32 crc kubenswrapper[4934]: I1002 10:00:32.385732 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ef8e1eb9-668c-4542-9551-8a02b13222b9-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-msxrz\" (UID: \"ef8e1eb9-668c-4542-9551-8a02b13222b9\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-msxrz" Oct 02 10:00:32 crc kubenswrapper[4934]: I1002 10:00:32.474119 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-w6z5d" Oct 02 10:00:32 crc kubenswrapper[4934]: I1002 10:00:32.500540 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7cc8bd7b4-jmbrc"] Oct 02 10:00:32 crc kubenswrapper[4934]: W1002 10:00:32.508074 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc1b9dffc_787b_4750_8242_e2716633ca40.slice/crio-0f880feefa5f0770a18fe955c8c6485deadc1d3fa1c7acff4430998eebd98e44 WatchSource:0}: Error finding container 0f880feefa5f0770a18fe955c8c6485deadc1d3fa1c7acff4430998eebd98e44: Status 404 returned error can't find the container with id 0f880feefa5f0770a18fe955c8c6485deadc1d3fa1c7acff4430998eebd98e44 Oct 02 10:00:32 crc kubenswrapper[4934]: I1002 10:00:32.510414 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-lwrzm" event={"ID":"d705acd3-4167-430e-b33d-5966e485a9ce","Type":"ContainerStarted","Data":"54737e83137266924ca8a77ff7fec67d351eef7ece5fe1576f51ab75f86ef95f"} Oct 02 10:00:32 crc kubenswrapper[4934]: I1002 10:00:32.511514 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-mb8mr" event={"ID":"687810e0-5141-42c9-a40b-5f04d045bd5d","Type":"ContainerStarted","Data":"689865ed5b92193654e0e1b62d329c8f763e773776319634149eef73aa3b6cbb"} Oct 02 10:00:32 crc kubenswrapper[4934]: I1002 10:00:32.631631 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-msxrz" Oct 02 10:00:32 crc kubenswrapper[4934]: I1002 10:00:32.707693 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-w6z5d"] Oct 02 10:00:32 crc kubenswrapper[4934]: I1002 10:00:32.858563 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-msxrz"] Oct 02 10:00:32 crc kubenswrapper[4934]: W1002 10:00:32.865517 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef8e1eb9_668c_4542_9551_8a02b13222b9.slice/crio-742a37b67922b25e574c5752cb7fc0df4c8e8805244b15e15f468550847e4fef WatchSource:0}: Error finding container 742a37b67922b25e574c5752cb7fc0df4c8e8805244b15e15f468550847e4fef: Status 404 returned error can't find the container with id 742a37b67922b25e574c5752cb7fc0df4c8e8805244b15e15f468550847e4fef Oct 02 10:00:33 crc kubenswrapper[4934]: I1002 10:00:33.517164 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-msxrz" event={"ID":"ef8e1eb9-668c-4542-9551-8a02b13222b9","Type":"ContainerStarted","Data":"742a37b67922b25e574c5752cb7fc0df4c8e8805244b15e15f468550847e4fef"} Oct 02 10:00:33 crc kubenswrapper[4934]: I1002 10:00:33.518522 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7cc8bd7b4-jmbrc" event={"ID":"c1b9dffc-787b-4750-8242-e2716633ca40","Type":"ContainerStarted","Data":"9954ee6398542c1f423cd4b01b474e1fe33b7a2dd8239c2e73c0dc6f6ebcab08"} Oct 02 10:00:33 crc kubenswrapper[4934]: I1002 10:00:33.518556 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7cc8bd7b4-jmbrc" event={"ID":"c1b9dffc-787b-4750-8242-e2716633ca40","Type":"ContainerStarted","Data":"0f880feefa5f0770a18fe955c8c6485deadc1d3fa1c7acff4430998eebd98e44"} Oct 02 10:00:33 crc kubenswrapper[4934]: I1002 10:00:33.519989 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-w6z5d" event={"ID":"f45d40d5-581b-47e0-a25e-9c550f92f689","Type":"ContainerStarted","Data":"959f0c62402105ee8012a399fe2ccc654fb33c416d46499348bf810cb6bdaf2e"} Oct 02 10:00:33 crc kubenswrapper[4934]: I1002 10:00:33.542288 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7cc8bd7b4-jmbrc" podStartSLOduration=2.5422673959999997 podStartE2EDuration="2.542267396s" podCreationTimestamp="2025-10-02 10:00:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:00:33.536528638 +0000 UTC m=+705.289170170" watchObservedRunningTime="2025-10-02 10:00:33.542267396 +0000 UTC m=+705.294908918" Oct 02 10:00:34 crc kubenswrapper[4934]: I1002 10:00:34.529361 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-mb8mr" event={"ID":"687810e0-5141-42c9-a40b-5f04d045bd5d","Type":"ContainerStarted","Data":"0c6e55dd752a315137aa775f78936387da2827b89568c0bb85375ccef537ba15"} Oct 02 10:00:34 crc kubenswrapper[4934]: I1002 10:00:34.529953 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-mb8mr" Oct 02 10:00:34 crc kubenswrapper[4934]: I1002 10:00:34.532186 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-lwrzm" event={"ID":"d705acd3-4167-430e-b33d-5966e485a9ce","Type":"ContainerStarted","Data":"a369afda844f187ba7fe4904c95924485da4a4256ed6fdc7e06c956f88375da5"} Oct 02 10:00:34 crc kubenswrapper[4934]: I1002 10:00:34.537939 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-w6z5d" event={"ID":"f45d40d5-581b-47e0-a25e-9c550f92f689","Type":"ContainerStarted","Data":"1478484697ee78df6b4fab622a85adffd9a4d5d695b2058cd4d24218d3ed3044"} Oct 02 10:00:34 crc kubenswrapper[4934]: I1002 10:00:34.538076 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-w6z5d" Oct 02 10:00:34 crc kubenswrapper[4934]: I1002 10:00:34.543461 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-mb8mr" podStartSLOduration=1.275509467 podStartE2EDuration="3.543443775s" podCreationTimestamp="2025-10-02 10:00:31 +0000 UTC" firstStartedPulling="2025-10-02 10:00:31.950939184 +0000 UTC m=+703.703580706" lastFinishedPulling="2025-10-02 10:00:34.218873492 +0000 UTC m=+705.971515014" observedRunningTime="2025-10-02 10:00:34.541802451 +0000 UTC m=+706.294443983" watchObservedRunningTime="2025-10-02 10:00:34.543443775 +0000 UTC m=+706.296085297" Oct 02 10:00:34 crc kubenswrapper[4934]: I1002 10:00:34.559841 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-w6z5d" podStartSLOduration=2.076910272 podStartE2EDuration="3.559817537s" podCreationTimestamp="2025-10-02 10:00:31 +0000 UTC" firstStartedPulling="2025-10-02 10:00:32.725811697 +0000 UTC m=+704.478453229" lastFinishedPulling="2025-10-02 10:00:34.208718972 +0000 UTC m=+705.961360494" observedRunningTime="2025-10-02 10:00:34.556373262 +0000 UTC m=+706.309014794" watchObservedRunningTime="2025-10-02 10:00:34.559817537 +0000 UTC m=+706.312459079" Oct 02 10:00:35 crc kubenswrapper[4934]: I1002 10:00:35.548216 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-msxrz" event={"ID":"ef8e1eb9-668c-4542-9551-8a02b13222b9","Type":"ContainerStarted","Data":"39ea5db397139dc11b8d7c7320c661cc63bf541d6a2bf0c727efeb28e9d6d4d7"} Oct 02 10:00:35 crc kubenswrapper[4934]: I1002 10:00:35.572639 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-msxrz" podStartSLOduration=2.227711008 podStartE2EDuration="4.572568156s" podCreationTimestamp="2025-10-02 10:00:31 +0000 UTC" firstStartedPulling="2025-10-02 10:00:32.867903372 +0000 UTC m=+704.620544894" lastFinishedPulling="2025-10-02 10:00:35.21276051 +0000 UTC m=+706.965402042" observedRunningTime="2025-10-02 10:00:35.567326391 +0000 UTC m=+707.319967933" watchObservedRunningTime="2025-10-02 10:00:35.572568156 +0000 UTC m=+707.325209718" Oct 02 10:00:37 crc kubenswrapper[4934]: I1002 10:00:37.562140 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-lwrzm" event={"ID":"d705acd3-4167-430e-b33d-5966e485a9ce","Type":"ContainerStarted","Data":"3f174fb0ffe9a062196a9eafa1c03aa4c769cfacd9e09fb4840f8342a539bfdd"} Oct 02 10:00:37 crc kubenswrapper[4934]: I1002 10:00:37.589148 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-lwrzm" podStartSLOduration=2.19220369 podStartE2EDuration="6.589116536s" podCreationTimestamp="2025-10-02 10:00:31 +0000 UTC" firstStartedPulling="2025-10-02 10:00:32.102702666 +0000 UTC m=+703.855344188" lastFinishedPulling="2025-10-02 10:00:36.499615512 +0000 UTC m=+708.252257034" observedRunningTime="2025-10-02 10:00:37.578087042 +0000 UTC m=+709.330728634" watchObservedRunningTime="2025-10-02 10:00:37.589116536 +0000 UTC m=+709.341758098" Oct 02 10:00:38 crc kubenswrapper[4934]: I1002 10:00:38.439345 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:00:38 crc kubenswrapper[4934]: I1002 10:00:38.439634 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:00:38 crc kubenswrapper[4934]: I1002 10:00:38.581895 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-kcgtd" Oct 02 10:00:41 crc kubenswrapper[4934]: I1002 10:00:41.947402 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-mb8mr" Oct 02 10:00:42 crc kubenswrapper[4934]: I1002 10:00:42.247371 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7cc8bd7b4-jmbrc" Oct 02 10:00:42 crc kubenswrapper[4934]: I1002 10:00:42.247447 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7cc8bd7b4-jmbrc" Oct 02 10:00:42 crc kubenswrapper[4934]: I1002 10:00:42.251508 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7cc8bd7b4-jmbrc" Oct 02 10:00:42 crc kubenswrapper[4934]: I1002 10:00:42.595418 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7cc8bd7b4-jmbrc" Oct 02 10:00:42 crc kubenswrapper[4934]: I1002 10:00:42.659776 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-f44r4"] Oct 02 10:00:49 crc kubenswrapper[4934]: I1002 10:00:49.294500 4934 scope.go:117] "RemoveContainer" containerID="08f4bbc53be5040dff8a07d5db8c29324bc0a6f209909d39639b66aec504fa65" Oct 02 10:00:49 crc kubenswrapper[4934]: I1002 10:00:49.643026 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-79fxg_73968247-67dd-48cc-88a1-64afac657412/kube-multus/1.log" Oct 02 10:00:52 crc kubenswrapper[4934]: I1002 10:00:52.480875 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-w6z5d" Oct 02 10:00:52 crc kubenswrapper[4934]: I1002 10:00:52.935814 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q2p9l"] Oct 02 10:00:52 crc kubenswrapper[4934]: I1002 10:00:52.936080 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-q2p9l" podUID="f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9" containerName="controller-manager" containerID="cri-o://6df8a7c62f24bef35b7570cf67499418e438e5978cf6864f33decbe9546a8d3e" gracePeriod=30 Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.037998 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6"] Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.038211 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6" podUID="87b8c00f-e910-42f0-8089-73df46c6cf0b" containerName="route-controller-manager" containerID="cri-o://1f42a757d9a24bdf290901c28bdbb637a8e44d2cdd551ec0573b9f29fa2d07f3" gracePeriod=30 Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.285963 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-q2p9l" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.357090 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-serving-cert\") pod \"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9\" (UID: \"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9\") " Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.357157 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-client-ca\") pod \"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9\" (UID: \"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9\") " Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.357189 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zn9w7\" (UniqueName: \"kubernetes.io/projected/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-kube-api-access-zn9w7\") pod \"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9\" (UID: \"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9\") " Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.357233 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-proxy-ca-bundles\") pod \"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9\" (UID: \"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9\") " Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.357285 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-config\") pod \"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9\" (UID: \"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9\") " Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.358537 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-config" (OuterVolumeSpecName: "config") pod "f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9" (UID: "f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.361800 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9" (UID: "f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.361853 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-client-ca" (OuterVolumeSpecName: "client-ca") pod "f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9" (UID: "f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.377901 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9" (UID: "f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.378384 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-kube-api-access-zn9w7" (OuterVolumeSpecName: "kube-api-access-zn9w7") pod "f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9" (UID: "f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9"). InnerVolumeSpecName "kube-api-access-zn9w7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.386809 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.459054 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/87b8c00f-e910-42f0-8089-73df46c6cf0b-client-ca\") pod \"87b8c00f-e910-42f0-8089-73df46c6cf0b\" (UID: \"87b8c00f-e910-42f0-8089-73df46c6cf0b\") " Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.459132 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87b8c00f-e910-42f0-8089-73df46c6cf0b-config\") pod \"87b8c00f-e910-42f0-8089-73df46c6cf0b\" (UID: \"87b8c00f-e910-42f0-8089-73df46c6cf0b\") " Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.459212 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svdvd\" (UniqueName: \"kubernetes.io/projected/87b8c00f-e910-42f0-8089-73df46c6cf0b-kube-api-access-svdvd\") pod \"87b8c00f-e910-42f0-8089-73df46c6cf0b\" (UID: \"87b8c00f-e910-42f0-8089-73df46c6cf0b\") " Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.459254 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87b8c00f-e910-42f0-8089-73df46c6cf0b-serving-cert\") pod \"87b8c00f-e910-42f0-8089-73df46c6cf0b\" (UID: \"87b8c00f-e910-42f0-8089-73df46c6cf0b\") " Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.460265 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.460291 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.460307 4934 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.460321 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zn9w7\" (UniqueName: \"kubernetes.io/projected/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-kube-api-access-zn9w7\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.460336 4934 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.460436 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87b8c00f-e910-42f0-8089-73df46c6cf0b-client-ca" (OuterVolumeSpecName: "client-ca") pod "87b8c00f-e910-42f0-8089-73df46c6cf0b" (UID: "87b8c00f-e910-42f0-8089-73df46c6cf0b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.460478 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87b8c00f-e910-42f0-8089-73df46c6cf0b-config" (OuterVolumeSpecName: "config") pod "87b8c00f-e910-42f0-8089-73df46c6cf0b" (UID: "87b8c00f-e910-42f0-8089-73df46c6cf0b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.463621 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87b8c00f-e910-42f0-8089-73df46c6cf0b-kube-api-access-svdvd" (OuterVolumeSpecName: "kube-api-access-svdvd") pod "87b8c00f-e910-42f0-8089-73df46c6cf0b" (UID: "87b8c00f-e910-42f0-8089-73df46c6cf0b"). InnerVolumeSpecName "kube-api-access-svdvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.463980 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87b8c00f-e910-42f0-8089-73df46c6cf0b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "87b8c00f-e910-42f0-8089-73df46c6cf0b" (UID: "87b8c00f-e910-42f0-8089-73df46c6cf0b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.561489 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svdvd\" (UniqueName: \"kubernetes.io/projected/87b8c00f-e910-42f0-8089-73df46c6cf0b-kube-api-access-svdvd\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.561546 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87b8c00f-e910-42f0-8089-73df46c6cf0b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.561568 4934 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/87b8c00f-e910-42f0-8089-73df46c6cf0b-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.561609 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87b8c00f-e910-42f0-8089-73df46c6cf0b-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.669366 4934 generic.go:334] "Generic (PLEG): container finished" podID="87b8c00f-e910-42f0-8089-73df46c6cf0b" containerID="1f42a757d9a24bdf290901c28bdbb637a8e44d2cdd551ec0573b9f29fa2d07f3" exitCode=0 Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.670165 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.670772 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6" event={"ID":"87b8c00f-e910-42f0-8089-73df46c6cf0b","Type":"ContainerDied","Data":"1f42a757d9a24bdf290901c28bdbb637a8e44d2cdd551ec0573b9f29fa2d07f3"} Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.670841 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6" event={"ID":"87b8c00f-e910-42f0-8089-73df46c6cf0b","Type":"ContainerDied","Data":"20b01acf128a5ec360f96e55587dd495ddd8b1ca26c0f2fe7c6932e2cdd82c4d"} Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.670862 4934 scope.go:117] "RemoveContainer" containerID="1f42a757d9a24bdf290901c28bdbb637a8e44d2cdd551ec0573b9f29fa2d07f3" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.673417 4934 generic.go:334] "Generic (PLEG): container finished" podID="f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9" containerID="6df8a7c62f24bef35b7570cf67499418e438e5978cf6864f33decbe9546a8d3e" exitCode=0 Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.673453 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-q2p9l" event={"ID":"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9","Type":"ContainerDied","Data":"6df8a7c62f24bef35b7570cf67499418e438e5978cf6864f33decbe9546a8d3e"} Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.673472 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-q2p9l" event={"ID":"f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9","Type":"ContainerDied","Data":"85bc11ec45ad9c47972b55d4674346fd4e8da517758ae63b529627b13151b7fa"} Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.673520 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-q2p9l" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.687813 4934 scope.go:117] "RemoveContainer" containerID="1f42a757d9a24bdf290901c28bdbb637a8e44d2cdd551ec0573b9f29fa2d07f3" Oct 02 10:00:53 crc kubenswrapper[4934]: E1002 10:00:53.688266 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f42a757d9a24bdf290901c28bdbb637a8e44d2cdd551ec0573b9f29fa2d07f3\": container with ID starting with 1f42a757d9a24bdf290901c28bdbb637a8e44d2cdd551ec0573b9f29fa2d07f3 not found: ID does not exist" containerID="1f42a757d9a24bdf290901c28bdbb637a8e44d2cdd551ec0573b9f29fa2d07f3" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.688334 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f42a757d9a24bdf290901c28bdbb637a8e44d2cdd551ec0573b9f29fa2d07f3"} err="failed to get container status \"1f42a757d9a24bdf290901c28bdbb637a8e44d2cdd551ec0573b9f29fa2d07f3\": rpc error: code = NotFound desc = could not find container \"1f42a757d9a24bdf290901c28bdbb637a8e44d2cdd551ec0573b9f29fa2d07f3\": container with ID starting with 1f42a757d9a24bdf290901c28bdbb637a8e44d2cdd551ec0573b9f29fa2d07f3 not found: ID does not exist" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.688365 4934 scope.go:117] "RemoveContainer" containerID="6df8a7c62f24bef35b7570cf67499418e438e5978cf6864f33decbe9546a8d3e" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.707792 4934 scope.go:117] "RemoveContainer" containerID="6df8a7c62f24bef35b7570cf67499418e438e5978cf6864f33decbe9546a8d3e" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.707878 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6"] Oct 02 10:00:53 crc kubenswrapper[4934]: E1002 10:00:53.708271 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6df8a7c62f24bef35b7570cf67499418e438e5978cf6864f33decbe9546a8d3e\": container with ID starting with 6df8a7c62f24bef35b7570cf67499418e438e5978cf6864f33decbe9546a8d3e not found: ID does not exist" containerID="6df8a7c62f24bef35b7570cf67499418e438e5978cf6864f33decbe9546a8d3e" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.708308 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6df8a7c62f24bef35b7570cf67499418e438e5978cf6864f33decbe9546a8d3e"} err="failed to get container status \"6df8a7c62f24bef35b7570cf67499418e438e5978cf6864f33decbe9546a8d3e\": rpc error: code = NotFound desc = could not find container \"6df8a7c62f24bef35b7570cf67499418e438e5978cf6864f33decbe9546a8d3e\": container with ID starting with 6df8a7c62f24bef35b7570cf67499418e438e5978cf6864f33decbe9546a8d3e not found: ID does not exist" Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.716907 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-d6cv6"] Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.720373 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q2p9l"] Oct 02 10:00:53 crc kubenswrapper[4934]: I1002 10:00:53.722856 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q2p9l"] Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.832737 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-788d87485d-rmkx2"] Oct 02 10:00:54 crc kubenswrapper[4934]: E1002 10:00:54.833011 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87b8c00f-e910-42f0-8089-73df46c6cf0b" containerName="route-controller-manager" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.833024 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="87b8c00f-e910-42f0-8089-73df46c6cf0b" containerName="route-controller-manager" Oct 02 10:00:54 crc kubenswrapper[4934]: E1002 10:00:54.833047 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9" containerName="controller-manager" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.833053 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9" containerName="controller-manager" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.833152 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9" containerName="controller-manager" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.833161 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="87b8c00f-e910-42f0-8089-73df46c6cf0b" containerName="route-controller-manager" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.833606 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-788d87485d-rmkx2" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.836067 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-57f66c84b5-btxzz"] Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.836860 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57f66c84b5-btxzz" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.838798 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.838976 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.839186 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.839396 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.839736 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.839967 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.840143 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.840256 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-788d87485d-rmkx2"] Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.842768 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.844096 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.844142 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.844194 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.844294 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.844591 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-57f66c84b5-btxzz"] Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.848728 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.877839 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwn2c\" (UniqueName: \"kubernetes.io/projected/cbdf1bf4-74c4-4ceb-9eee-8fae1a423940-kube-api-access-mwn2c\") pod \"controller-manager-57f66c84b5-btxzz\" (UID: \"cbdf1bf4-74c4-4ceb-9eee-8fae1a423940\") " pod="openshift-controller-manager/controller-manager-57f66c84b5-btxzz" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.878181 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbdf1bf4-74c4-4ceb-9eee-8fae1a423940-serving-cert\") pod \"controller-manager-57f66c84b5-btxzz\" (UID: \"cbdf1bf4-74c4-4ceb-9eee-8fae1a423940\") " pod="openshift-controller-manager/controller-manager-57f66c84b5-btxzz" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.878206 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbdf1bf4-74c4-4ceb-9eee-8fae1a423940-config\") pod \"controller-manager-57f66c84b5-btxzz\" (UID: \"cbdf1bf4-74c4-4ceb-9eee-8fae1a423940\") " pod="openshift-controller-manager/controller-manager-57f66c84b5-btxzz" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.878249 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cbdf1bf4-74c4-4ceb-9eee-8fae1a423940-client-ca\") pod \"controller-manager-57f66c84b5-btxzz\" (UID: \"cbdf1bf4-74c4-4ceb-9eee-8fae1a423940\") " pod="openshift-controller-manager/controller-manager-57f66c84b5-btxzz" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.878266 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v92gt\" (UniqueName: \"kubernetes.io/projected/f20a160b-61ca-4ce6-a027-b0d715e5985a-kube-api-access-v92gt\") pod \"route-controller-manager-788d87485d-rmkx2\" (UID: \"f20a160b-61ca-4ce6-a027-b0d715e5985a\") " pod="openshift-route-controller-manager/route-controller-manager-788d87485d-rmkx2" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.878433 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cbdf1bf4-74c4-4ceb-9eee-8fae1a423940-proxy-ca-bundles\") pod \"controller-manager-57f66c84b5-btxzz\" (UID: \"cbdf1bf4-74c4-4ceb-9eee-8fae1a423940\") " pod="openshift-controller-manager/controller-manager-57f66c84b5-btxzz" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.878504 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f20a160b-61ca-4ce6-a027-b0d715e5985a-client-ca\") pod \"route-controller-manager-788d87485d-rmkx2\" (UID: \"f20a160b-61ca-4ce6-a027-b0d715e5985a\") " pod="openshift-route-controller-manager/route-controller-manager-788d87485d-rmkx2" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.878568 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f20a160b-61ca-4ce6-a027-b0d715e5985a-serving-cert\") pod \"route-controller-manager-788d87485d-rmkx2\" (UID: \"f20a160b-61ca-4ce6-a027-b0d715e5985a\") " pod="openshift-route-controller-manager/route-controller-manager-788d87485d-rmkx2" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.878654 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f20a160b-61ca-4ce6-a027-b0d715e5985a-config\") pod \"route-controller-manager-788d87485d-rmkx2\" (UID: \"f20a160b-61ca-4ce6-a027-b0d715e5985a\") " pod="openshift-route-controller-manager/route-controller-manager-788d87485d-rmkx2" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.920764 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87b8c00f-e910-42f0-8089-73df46c6cf0b" path="/var/lib/kubelet/pods/87b8c00f-e910-42f0-8089-73df46c6cf0b/volumes" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.921672 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9" path="/var/lib/kubelet/pods/f9e241d8-3d16-4cce-bfec-46d4b3bcb7b9/volumes" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.941606 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-788d87485d-rmkx2"] Oct 02 10:00:54 crc kubenswrapper[4934]: E1002 10:00:54.942055 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[client-ca config kube-api-access-v92gt serving-cert], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-route-controller-manager/route-controller-manager-788d87485d-rmkx2" podUID="f20a160b-61ca-4ce6-a027-b0d715e5985a" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.979727 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbdf1bf4-74c4-4ceb-9eee-8fae1a423940-config\") pod \"controller-manager-57f66c84b5-btxzz\" (UID: \"cbdf1bf4-74c4-4ceb-9eee-8fae1a423940\") " pod="openshift-controller-manager/controller-manager-57f66c84b5-btxzz" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.979812 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cbdf1bf4-74c4-4ceb-9eee-8fae1a423940-client-ca\") pod \"controller-manager-57f66c84b5-btxzz\" (UID: \"cbdf1bf4-74c4-4ceb-9eee-8fae1a423940\") " pod="openshift-controller-manager/controller-manager-57f66c84b5-btxzz" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.979835 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v92gt\" (UniqueName: \"kubernetes.io/projected/f20a160b-61ca-4ce6-a027-b0d715e5985a-kube-api-access-v92gt\") pod \"route-controller-manager-788d87485d-rmkx2\" (UID: \"f20a160b-61ca-4ce6-a027-b0d715e5985a\") " pod="openshift-route-controller-manager/route-controller-manager-788d87485d-rmkx2" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.979892 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cbdf1bf4-74c4-4ceb-9eee-8fae1a423940-proxy-ca-bundles\") pod \"controller-manager-57f66c84b5-btxzz\" (UID: \"cbdf1bf4-74c4-4ceb-9eee-8fae1a423940\") " pod="openshift-controller-manager/controller-manager-57f66c84b5-btxzz" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.979907 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f20a160b-61ca-4ce6-a027-b0d715e5985a-client-ca\") pod \"route-controller-manager-788d87485d-rmkx2\" (UID: \"f20a160b-61ca-4ce6-a027-b0d715e5985a\") " pod="openshift-route-controller-manager/route-controller-manager-788d87485d-rmkx2" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.979964 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f20a160b-61ca-4ce6-a027-b0d715e5985a-serving-cert\") pod \"route-controller-manager-788d87485d-rmkx2\" (UID: \"f20a160b-61ca-4ce6-a027-b0d715e5985a\") " pod="openshift-route-controller-manager/route-controller-manager-788d87485d-rmkx2" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.979987 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f20a160b-61ca-4ce6-a027-b0d715e5985a-config\") pod \"route-controller-manager-788d87485d-rmkx2\" (UID: \"f20a160b-61ca-4ce6-a027-b0d715e5985a\") " pod="openshift-route-controller-manager/route-controller-manager-788d87485d-rmkx2" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.980006 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwn2c\" (UniqueName: \"kubernetes.io/projected/cbdf1bf4-74c4-4ceb-9eee-8fae1a423940-kube-api-access-mwn2c\") pod \"controller-manager-57f66c84b5-btxzz\" (UID: \"cbdf1bf4-74c4-4ceb-9eee-8fae1a423940\") " pod="openshift-controller-manager/controller-manager-57f66c84b5-btxzz" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.980025 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbdf1bf4-74c4-4ceb-9eee-8fae1a423940-serving-cert\") pod \"controller-manager-57f66c84b5-btxzz\" (UID: \"cbdf1bf4-74c4-4ceb-9eee-8fae1a423940\") " pod="openshift-controller-manager/controller-manager-57f66c84b5-btxzz" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.981709 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cbdf1bf4-74c4-4ceb-9eee-8fae1a423940-client-ca\") pod \"controller-manager-57f66c84b5-btxzz\" (UID: \"cbdf1bf4-74c4-4ceb-9eee-8fae1a423940\") " pod="openshift-controller-manager/controller-manager-57f66c84b5-btxzz" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.982488 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbdf1bf4-74c4-4ceb-9eee-8fae1a423940-config\") pod \"controller-manager-57f66c84b5-btxzz\" (UID: \"cbdf1bf4-74c4-4ceb-9eee-8fae1a423940\") " pod="openshift-controller-manager/controller-manager-57f66c84b5-btxzz" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.982669 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f20a160b-61ca-4ce6-a027-b0d715e5985a-client-ca\") pod \"route-controller-manager-788d87485d-rmkx2\" (UID: \"f20a160b-61ca-4ce6-a027-b0d715e5985a\") " pod="openshift-route-controller-manager/route-controller-manager-788d87485d-rmkx2" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.982842 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cbdf1bf4-74c4-4ceb-9eee-8fae1a423940-proxy-ca-bundles\") pod \"controller-manager-57f66c84b5-btxzz\" (UID: \"cbdf1bf4-74c4-4ceb-9eee-8fae1a423940\") " pod="openshift-controller-manager/controller-manager-57f66c84b5-btxzz" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.983331 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f20a160b-61ca-4ce6-a027-b0d715e5985a-config\") pod \"route-controller-manager-788d87485d-rmkx2\" (UID: \"f20a160b-61ca-4ce6-a027-b0d715e5985a\") " pod="openshift-route-controller-manager/route-controller-manager-788d87485d-rmkx2" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.986822 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbdf1bf4-74c4-4ceb-9eee-8fae1a423940-serving-cert\") pod \"controller-manager-57f66c84b5-btxzz\" (UID: \"cbdf1bf4-74c4-4ceb-9eee-8fae1a423940\") " pod="openshift-controller-manager/controller-manager-57f66c84b5-btxzz" Oct 02 10:00:54 crc kubenswrapper[4934]: I1002 10:00:54.987382 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f20a160b-61ca-4ce6-a027-b0d715e5985a-serving-cert\") pod \"route-controller-manager-788d87485d-rmkx2\" (UID: \"f20a160b-61ca-4ce6-a027-b0d715e5985a\") " pod="openshift-route-controller-manager/route-controller-manager-788d87485d-rmkx2" Oct 02 10:00:55 crc kubenswrapper[4934]: I1002 10:00:55.002556 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwn2c\" (UniqueName: \"kubernetes.io/projected/cbdf1bf4-74c4-4ceb-9eee-8fae1a423940-kube-api-access-mwn2c\") pod \"controller-manager-57f66c84b5-btxzz\" (UID: \"cbdf1bf4-74c4-4ceb-9eee-8fae1a423940\") " pod="openshift-controller-manager/controller-manager-57f66c84b5-btxzz" Oct 02 10:00:55 crc kubenswrapper[4934]: I1002 10:00:55.010753 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v92gt\" (UniqueName: \"kubernetes.io/projected/f20a160b-61ca-4ce6-a027-b0d715e5985a-kube-api-access-v92gt\") pod \"route-controller-manager-788d87485d-rmkx2\" (UID: \"f20a160b-61ca-4ce6-a027-b0d715e5985a\") " pod="openshift-route-controller-manager/route-controller-manager-788d87485d-rmkx2" Oct 02 10:00:55 crc kubenswrapper[4934]: I1002 10:00:55.174908 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57f66c84b5-btxzz" Oct 02 10:00:55 crc kubenswrapper[4934]: I1002 10:00:55.628104 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-57f66c84b5-btxzz"] Oct 02 10:00:55 crc kubenswrapper[4934]: W1002 10:00:55.636163 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcbdf1bf4_74c4_4ceb_9eee_8fae1a423940.slice/crio-132a3efaad78b9c9b264013af99cba98fd2d6db9cb698b1ab71a23c409ce90ad WatchSource:0}: Error finding container 132a3efaad78b9c9b264013af99cba98fd2d6db9cb698b1ab71a23c409ce90ad: Status 404 returned error can't find the container with id 132a3efaad78b9c9b264013af99cba98fd2d6db9cb698b1ab71a23c409ce90ad Oct 02 10:00:55 crc kubenswrapper[4934]: I1002 10:00:55.686482 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-788d87485d-rmkx2" Oct 02 10:00:55 crc kubenswrapper[4934]: I1002 10:00:55.686472 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57f66c84b5-btxzz" event={"ID":"cbdf1bf4-74c4-4ceb-9eee-8fae1a423940","Type":"ContainerStarted","Data":"132a3efaad78b9c9b264013af99cba98fd2d6db9cb698b1ab71a23c409ce90ad"} Oct 02 10:00:55 crc kubenswrapper[4934]: I1002 10:00:55.716784 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-788d87485d-rmkx2" Oct 02 10:00:55 crc kubenswrapper[4934]: I1002 10:00:55.787851 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v92gt\" (UniqueName: \"kubernetes.io/projected/f20a160b-61ca-4ce6-a027-b0d715e5985a-kube-api-access-v92gt\") pod \"f20a160b-61ca-4ce6-a027-b0d715e5985a\" (UID: \"f20a160b-61ca-4ce6-a027-b0d715e5985a\") " Oct 02 10:00:55 crc kubenswrapper[4934]: I1002 10:00:55.787957 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f20a160b-61ca-4ce6-a027-b0d715e5985a-serving-cert\") pod \"f20a160b-61ca-4ce6-a027-b0d715e5985a\" (UID: \"f20a160b-61ca-4ce6-a027-b0d715e5985a\") " Oct 02 10:00:55 crc kubenswrapper[4934]: I1002 10:00:55.788009 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f20a160b-61ca-4ce6-a027-b0d715e5985a-config\") pod \"f20a160b-61ca-4ce6-a027-b0d715e5985a\" (UID: \"f20a160b-61ca-4ce6-a027-b0d715e5985a\") " Oct 02 10:00:55 crc kubenswrapper[4934]: I1002 10:00:55.788023 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f20a160b-61ca-4ce6-a027-b0d715e5985a-client-ca\") pod \"f20a160b-61ca-4ce6-a027-b0d715e5985a\" (UID: \"f20a160b-61ca-4ce6-a027-b0d715e5985a\") " Oct 02 10:00:55 crc kubenswrapper[4934]: I1002 10:00:55.788628 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f20a160b-61ca-4ce6-a027-b0d715e5985a-client-ca" (OuterVolumeSpecName: "client-ca") pod "f20a160b-61ca-4ce6-a027-b0d715e5985a" (UID: "f20a160b-61ca-4ce6-a027-b0d715e5985a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:55 crc kubenswrapper[4934]: I1002 10:00:55.788908 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f20a160b-61ca-4ce6-a027-b0d715e5985a-config" (OuterVolumeSpecName: "config") pod "f20a160b-61ca-4ce6-a027-b0d715e5985a" (UID: "f20a160b-61ca-4ce6-a027-b0d715e5985a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:55 crc kubenswrapper[4934]: I1002 10:00:55.794743 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f20a160b-61ca-4ce6-a027-b0d715e5985a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f20a160b-61ca-4ce6-a027-b0d715e5985a" (UID: "f20a160b-61ca-4ce6-a027-b0d715e5985a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:55 crc kubenswrapper[4934]: I1002 10:00:55.794845 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f20a160b-61ca-4ce6-a027-b0d715e5985a-kube-api-access-v92gt" (OuterVolumeSpecName: "kube-api-access-v92gt") pod "f20a160b-61ca-4ce6-a027-b0d715e5985a" (UID: "f20a160b-61ca-4ce6-a027-b0d715e5985a"). InnerVolumeSpecName "kube-api-access-v92gt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:00:55 crc kubenswrapper[4934]: I1002 10:00:55.889158 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v92gt\" (UniqueName: \"kubernetes.io/projected/f20a160b-61ca-4ce6-a027-b0d715e5985a-kube-api-access-v92gt\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:55 crc kubenswrapper[4934]: I1002 10:00:55.889220 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f20a160b-61ca-4ce6-a027-b0d715e5985a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:55 crc kubenswrapper[4934]: I1002 10:00:55.889231 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f20a160b-61ca-4ce6-a027-b0d715e5985a-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:55 crc kubenswrapper[4934]: I1002 10:00:55.889241 4934 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f20a160b-61ca-4ce6-a027-b0d715e5985a-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:56 crc kubenswrapper[4934]: I1002 10:00:56.692979 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-788d87485d-rmkx2" Oct 02 10:00:56 crc kubenswrapper[4934]: I1002 10:00:56.693110 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57f66c84b5-btxzz" event={"ID":"cbdf1bf4-74c4-4ceb-9eee-8fae1a423940","Type":"ContainerStarted","Data":"fdba241caeae99dbcf38995966f31db7835492d47e5db089104b8b0589dfeff7"} Oct 02 10:00:56 crc kubenswrapper[4934]: I1002 10:00:56.693477 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-57f66c84b5-btxzz" Oct 02 10:00:56 crc kubenswrapper[4934]: I1002 10:00:56.698612 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-57f66c84b5-btxzz" Oct 02 10:00:56 crc kubenswrapper[4934]: I1002 10:00:56.714830 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-57f66c84b5-btxzz" podStartSLOduration=3.714810838 podStartE2EDuration="3.714810838s" podCreationTimestamp="2025-10-02 10:00:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:00:56.71160302 +0000 UTC m=+728.464244542" watchObservedRunningTime="2025-10-02 10:00:56.714810838 +0000 UTC m=+728.467452380" Oct 02 10:00:56 crc kubenswrapper[4934]: I1002 10:00:56.775819 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74b95dd9bf-bk5x4"] Oct 02 10:00:56 crc kubenswrapper[4934]: I1002 10:00:56.776509 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-74b95dd9bf-bk5x4" Oct 02 10:00:56 crc kubenswrapper[4934]: I1002 10:00:56.780099 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-788d87485d-rmkx2"] Oct 02 10:00:56 crc kubenswrapper[4934]: I1002 10:00:56.781539 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 02 10:00:56 crc kubenswrapper[4934]: I1002 10:00:56.781914 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 02 10:00:56 crc kubenswrapper[4934]: I1002 10:00:56.782094 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 02 10:00:56 crc kubenswrapper[4934]: I1002 10:00:56.782237 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 02 10:00:56 crc kubenswrapper[4934]: I1002 10:00:56.782411 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 02 10:00:56 crc kubenswrapper[4934]: I1002 10:00:56.782555 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 02 10:00:56 crc kubenswrapper[4934]: I1002 10:00:56.784320 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74b95dd9bf-bk5x4"] Oct 02 10:00:56 crc kubenswrapper[4934]: I1002 10:00:56.787677 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-788d87485d-rmkx2"] Oct 02 10:00:56 crc kubenswrapper[4934]: I1002 10:00:56.899261 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eaeeccf7-7315-4d6d-bfd1-ad5688e24796-client-ca\") pod \"route-controller-manager-74b95dd9bf-bk5x4\" (UID: \"eaeeccf7-7315-4d6d-bfd1-ad5688e24796\") " pod="openshift-route-controller-manager/route-controller-manager-74b95dd9bf-bk5x4" Oct 02 10:00:56 crc kubenswrapper[4934]: I1002 10:00:56.899314 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eaeeccf7-7315-4d6d-bfd1-ad5688e24796-config\") pod \"route-controller-manager-74b95dd9bf-bk5x4\" (UID: \"eaeeccf7-7315-4d6d-bfd1-ad5688e24796\") " pod="openshift-route-controller-manager/route-controller-manager-74b95dd9bf-bk5x4" Oct 02 10:00:56 crc kubenswrapper[4934]: I1002 10:00:56.899350 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5pgp\" (UniqueName: \"kubernetes.io/projected/eaeeccf7-7315-4d6d-bfd1-ad5688e24796-kube-api-access-v5pgp\") pod \"route-controller-manager-74b95dd9bf-bk5x4\" (UID: \"eaeeccf7-7315-4d6d-bfd1-ad5688e24796\") " pod="openshift-route-controller-manager/route-controller-manager-74b95dd9bf-bk5x4" Oct 02 10:00:56 crc kubenswrapper[4934]: I1002 10:00:56.899910 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eaeeccf7-7315-4d6d-bfd1-ad5688e24796-serving-cert\") pod \"route-controller-manager-74b95dd9bf-bk5x4\" (UID: \"eaeeccf7-7315-4d6d-bfd1-ad5688e24796\") " pod="openshift-route-controller-manager/route-controller-manager-74b95dd9bf-bk5x4" Oct 02 10:00:56 crc kubenswrapper[4934]: I1002 10:00:56.937785 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f20a160b-61ca-4ce6-a027-b0d715e5985a" path="/var/lib/kubelet/pods/f20a160b-61ca-4ce6-a027-b0d715e5985a/volumes" Oct 02 10:00:57 crc kubenswrapper[4934]: I1002 10:00:57.001719 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eaeeccf7-7315-4d6d-bfd1-ad5688e24796-serving-cert\") pod \"route-controller-manager-74b95dd9bf-bk5x4\" (UID: \"eaeeccf7-7315-4d6d-bfd1-ad5688e24796\") " pod="openshift-route-controller-manager/route-controller-manager-74b95dd9bf-bk5x4" Oct 02 10:00:57 crc kubenswrapper[4934]: I1002 10:00:57.001833 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eaeeccf7-7315-4d6d-bfd1-ad5688e24796-client-ca\") pod \"route-controller-manager-74b95dd9bf-bk5x4\" (UID: \"eaeeccf7-7315-4d6d-bfd1-ad5688e24796\") " pod="openshift-route-controller-manager/route-controller-manager-74b95dd9bf-bk5x4" Oct 02 10:00:57 crc kubenswrapper[4934]: I1002 10:00:57.001859 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eaeeccf7-7315-4d6d-bfd1-ad5688e24796-config\") pod \"route-controller-manager-74b95dd9bf-bk5x4\" (UID: \"eaeeccf7-7315-4d6d-bfd1-ad5688e24796\") " pod="openshift-route-controller-manager/route-controller-manager-74b95dd9bf-bk5x4" Oct 02 10:00:57 crc kubenswrapper[4934]: I1002 10:00:57.002163 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5pgp\" (UniqueName: \"kubernetes.io/projected/eaeeccf7-7315-4d6d-bfd1-ad5688e24796-kube-api-access-v5pgp\") pod \"route-controller-manager-74b95dd9bf-bk5x4\" (UID: \"eaeeccf7-7315-4d6d-bfd1-ad5688e24796\") " pod="openshift-route-controller-manager/route-controller-manager-74b95dd9bf-bk5x4" Oct 02 10:00:57 crc kubenswrapper[4934]: I1002 10:00:57.003663 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eaeeccf7-7315-4d6d-bfd1-ad5688e24796-client-ca\") pod \"route-controller-manager-74b95dd9bf-bk5x4\" (UID: \"eaeeccf7-7315-4d6d-bfd1-ad5688e24796\") " pod="openshift-route-controller-manager/route-controller-manager-74b95dd9bf-bk5x4" Oct 02 10:00:57 crc kubenswrapper[4934]: I1002 10:00:57.004012 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eaeeccf7-7315-4d6d-bfd1-ad5688e24796-config\") pod \"route-controller-manager-74b95dd9bf-bk5x4\" (UID: \"eaeeccf7-7315-4d6d-bfd1-ad5688e24796\") " pod="openshift-route-controller-manager/route-controller-manager-74b95dd9bf-bk5x4" Oct 02 10:00:57 crc kubenswrapper[4934]: I1002 10:00:57.009502 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eaeeccf7-7315-4d6d-bfd1-ad5688e24796-serving-cert\") pod \"route-controller-manager-74b95dd9bf-bk5x4\" (UID: \"eaeeccf7-7315-4d6d-bfd1-ad5688e24796\") " pod="openshift-route-controller-manager/route-controller-manager-74b95dd9bf-bk5x4" Oct 02 10:00:57 crc kubenswrapper[4934]: I1002 10:00:57.023275 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5pgp\" (UniqueName: \"kubernetes.io/projected/eaeeccf7-7315-4d6d-bfd1-ad5688e24796-kube-api-access-v5pgp\") pod \"route-controller-manager-74b95dd9bf-bk5x4\" (UID: \"eaeeccf7-7315-4d6d-bfd1-ad5688e24796\") " pod="openshift-route-controller-manager/route-controller-manager-74b95dd9bf-bk5x4" Oct 02 10:00:57 crc kubenswrapper[4934]: I1002 10:00:57.096943 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-74b95dd9bf-bk5x4" Oct 02 10:00:57 crc kubenswrapper[4934]: I1002 10:00:57.304751 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-74b95dd9bf-bk5x4"] Oct 02 10:00:57 crc kubenswrapper[4934]: I1002 10:00:57.701228 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-74b95dd9bf-bk5x4" event={"ID":"eaeeccf7-7315-4d6d-bfd1-ad5688e24796","Type":"ContainerStarted","Data":"e04ed2e4ed80b0700b959ae0f8fe7b6181bce8d4fec0eaa9552777fe4a902c6a"} Oct 02 10:00:57 crc kubenswrapper[4934]: I1002 10:00:57.701303 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-74b95dd9bf-bk5x4" event={"ID":"eaeeccf7-7315-4d6d-bfd1-ad5688e24796","Type":"ContainerStarted","Data":"6e83b5dd729405654b0fa160dc789ca7028ba24668f839b86cfcb7a35f33d78a"} Oct 02 10:00:57 crc kubenswrapper[4934]: I1002 10:00:57.701689 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-74b95dd9bf-bk5x4" Oct 02 10:00:57 crc kubenswrapper[4934]: I1002 10:00:57.719862 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-74b95dd9bf-bk5x4" podStartSLOduration=3.719841194 podStartE2EDuration="3.719841194s" podCreationTimestamp="2025-10-02 10:00:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:00:57.716528142 +0000 UTC m=+729.469169654" watchObservedRunningTime="2025-10-02 10:00:57.719841194 +0000 UTC m=+729.472482716" Oct 02 10:00:57 crc kubenswrapper[4934]: I1002 10:00:57.861371 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-74b95dd9bf-bk5x4" Oct 02 10:01:02 crc kubenswrapper[4934]: I1002 10:01:02.035763 4934 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 10:01:04 crc kubenswrapper[4934]: I1002 10:01:04.674782 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7qdvp"] Oct 02 10:01:04 crc kubenswrapper[4934]: I1002 10:01:04.677136 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7qdvp" Oct 02 10:01:04 crc kubenswrapper[4934]: I1002 10:01:04.692850 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7qdvp"] Oct 02 10:01:04 crc kubenswrapper[4934]: I1002 10:01:04.715532 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1888f99-8262-4f0e-97b8-fabef87b7557-catalog-content\") pod \"community-operators-7qdvp\" (UID: \"c1888f99-8262-4f0e-97b8-fabef87b7557\") " pod="openshift-marketplace/community-operators-7qdvp" Oct 02 10:01:04 crc kubenswrapper[4934]: I1002 10:01:04.715664 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1888f99-8262-4f0e-97b8-fabef87b7557-utilities\") pod \"community-operators-7qdvp\" (UID: \"c1888f99-8262-4f0e-97b8-fabef87b7557\") " pod="openshift-marketplace/community-operators-7qdvp" Oct 02 10:01:04 crc kubenswrapper[4934]: I1002 10:01:04.715728 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgbm9\" (UniqueName: \"kubernetes.io/projected/c1888f99-8262-4f0e-97b8-fabef87b7557-kube-api-access-tgbm9\") pod \"community-operators-7qdvp\" (UID: \"c1888f99-8262-4f0e-97b8-fabef87b7557\") " pod="openshift-marketplace/community-operators-7qdvp" Oct 02 10:01:04 crc kubenswrapper[4934]: I1002 10:01:04.817108 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1888f99-8262-4f0e-97b8-fabef87b7557-catalog-content\") pod \"community-operators-7qdvp\" (UID: \"c1888f99-8262-4f0e-97b8-fabef87b7557\") " pod="openshift-marketplace/community-operators-7qdvp" Oct 02 10:01:04 crc kubenswrapper[4934]: I1002 10:01:04.817199 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1888f99-8262-4f0e-97b8-fabef87b7557-utilities\") pod \"community-operators-7qdvp\" (UID: \"c1888f99-8262-4f0e-97b8-fabef87b7557\") " pod="openshift-marketplace/community-operators-7qdvp" Oct 02 10:01:04 crc kubenswrapper[4934]: I1002 10:01:04.817266 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgbm9\" (UniqueName: \"kubernetes.io/projected/c1888f99-8262-4f0e-97b8-fabef87b7557-kube-api-access-tgbm9\") pod \"community-operators-7qdvp\" (UID: \"c1888f99-8262-4f0e-97b8-fabef87b7557\") " pod="openshift-marketplace/community-operators-7qdvp" Oct 02 10:01:04 crc kubenswrapper[4934]: I1002 10:01:04.818651 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1888f99-8262-4f0e-97b8-fabef87b7557-catalog-content\") pod \"community-operators-7qdvp\" (UID: \"c1888f99-8262-4f0e-97b8-fabef87b7557\") " pod="openshift-marketplace/community-operators-7qdvp" Oct 02 10:01:04 crc kubenswrapper[4934]: I1002 10:01:04.818942 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1888f99-8262-4f0e-97b8-fabef87b7557-utilities\") pod \"community-operators-7qdvp\" (UID: \"c1888f99-8262-4f0e-97b8-fabef87b7557\") " pod="openshift-marketplace/community-operators-7qdvp" Oct 02 10:01:04 crc kubenswrapper[4934]: I1002 10:01:04.843114 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgbm9\" (UniqueName: \"kubernetes.io/projected/c1888f99-8262-4f0e-97b8-fabef87b7557-kube-api-access-tgbm9\") pod \"community-operators-7qdvp\" (UID: \"c1888f99-8262-4f0e-97b8-fabef87b7557\") " pod="openshift-marketplace/community-operators-7qdvp" Oct 02 10:01:04 crc kubenswrapper[4934]: I1002 10:01:04.997818 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7qdvp" Oct 02 10:01:05 crc kubenswrapper[4934]: I1002 10:01:05.487901 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7qdvp"] Oct 02 10:01:05 crc kubenswrapper[4934]: I1002 10:01:05.761387 4934 generic.go:334] "Generic (PLEG): container finished" podID="c1888f99-8262-4f0e-97b8-fabef87b7557" containerID="b60fc3bdbc4000acc6f1a506cb9d6e781c7dd640c8e449f619a13bb77d68164f" exitCode=0 Oct 02 10:01:05 crc kubenswrapper[4934]: I1002 10:01:05.761441 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7qdvp" event={"ID":"c1888f99-8262-4f0e-97b8-fabef87b7557","Type":"ContainerDied","Data":"b60fc3bdbc4000acc6f1a506cb9d6e781c7dd640c8e449f619a13bb77d68164f"} Oct 02 10:01:05 crc kubenswrapper[4934]: I1002 10:01:05.761478 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7qdvp" event={"ID":"c1888f99-8262-4f0e-97b8-fabef87b7557","Type":"ContainerStarted","Data":"a54efa35640922cecc7d46059ae35a2b478d75fb4b6191bb56e7aaade89e268f"} Oct 02 10:01:06 crc kubenswrapper[4934]: I1002 10:01:06.769027 4934 generic.go:334] "Generic (PLEG): container finished" podID="c1888f99-8262-4f0e-97b8-fabef87b7557" containerID="d1bfd10deb3932845a69155577eef144e9d3c8d6629ed7427562ff7e81779e4e" exitCode=0 Oct 02 10:01:06 crc kubenswrapper[4934]: I1002 10:01:06.769120 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7qdvp" event={"ID":"c1888f99-8262-4f0e-97b8-fabef87b7557","Type":"ContainerDied","Data":"d1bfd10deb3932845a69155577eef144e9d3c8d6629ed7427562ff7e81779e4e"} Oct 02 10:01:07 crc kubenswrapper[4934]: I1002 10:01:07.699337 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-f44r4" podUID="ad155aa0-50a3-4874-93bd-2166a8d093cb" containerName="console" containerID="cri-o://4a7f374112692e2fb8e8ed98087537f84a98f191b103da6bc4f4be2c2d11d5cc" gracePeriod=15 Oct 02 10:01:07 crc kubenswrapper[4934]: I1002 10:01:07.776235 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7qdvp" event={"ID":"c1888f99-8262-4f0e-97b8-fabef87b7557","Type":"ContainerStarted","Data":"d44f40b0a05d6078cca95175e8a987d65d9c7ebf9cf3db408d4a1f58c86988b5"} Oct 02 10:01:07 crc kubenswrapper[4934]: I1002 10:01:07.793859 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7qdvp" podStartSLOduration=2.223213883 podStartE2EDuration="3.793841116s" podCreationTimestamp="2025-10-02 10:01:04 +0000 UTC" firstStartedPulling="2025-10-02 10:01:05.764218814 +0000 UTC m=+737.516860336" lastFinishedPulling="2025-10-02 10:01:07.334846047 +0000 UTC m=+739.087487569" observedRunningTime="2025-10-02 10:01:07.790958567 +0000 UTC m=+739.543600099" watchObservedRunningTime="2025-10-02 10:01:07.793841116 +0000 UTC m=+739.546482648" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.204536 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-f44r4_ad155aa0-50a3-4874-93bd-2166a8d093cb/console/0.log" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.204609 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-f44r4" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.374442 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kr89h\" (UniqueName: \"kubernetes.io/projected/ad155aa0-50a3-4874-93bd-2166a8d093cb-kube-api-access-kr89h\") pod \"ad155aa0-50a3-4874-93bd-2166a8d093cb\" (UID: \"ad155aa0-50a3-4874-93bd-2166a8d093cb\") " Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.374740 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ad155aa0-50a3-4874-93bd-2166a8d093cb-service-ca\") pod \"ad155aa0-50a3-4874-93bd-2166a8d093cb\" (UID: \"ad155aa0-50a3-4874-93bd-2166a8d093cb\") " Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.374880 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ad155aa0-50a3-4874-93bd-2166a8d093cb-console-oauth-config\") pod \"ad155aa0-50a3-4874-93bd-2166a8d093cb\" (UID: \"ad155aa0-50a3-4874-93bd-2166a8d093cb\") " Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.375531 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad155aa0-50a3-4874-93bd-2166a8d093cb-service-ca" (OuterVolumeSpecName: "service-ca") pod "ad155aa0-50a3-4874-93bd-2166a8d093cb" (UID: "ad155aa0-50a3-4874-93bd-2166a8d093cb"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.376009 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ad155aa0-50a3-4874-93bd-2166a8d093cb-console-config\") pod \"ad155aa0-50a3-4874-93bd-2166a8d093cb\" (UID: \"ad155aa0-50a3-4874-93bd-2166a8d093cb\") " Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.376039 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad155aa0-50a3-4874-93bd-2166a8d093cb-trusted-ca-bundle\") pod \"ad155aa0-50a3-4874-93bd-2166a8d093cb\" (UID: \"ad155aa0-50a3-4874-93bd-2166a8d093cb\") " Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.376453 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad155aa0-50a3-4874-93bd-2166a8d093cb-console-config" (OuterVolumeSpecName: "console-config") pod "ad155aa0-50a3-4874-93bd-2166a8d093cb" (UID: "ad155aa0-50a3-4874-93bd-2166a8d093cb"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.376519 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ad155aa0-50a3-4874-93bd-2166a8d093cb-console-serving-cert\") pod \"ad155aa0-50a3-4874-93bd-2166a8d093cb\" (UID: \"ad155aa0-50a3-4874-93bd-2166a8d093cb\") " Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.376712 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad155aa0-50a3-4874-93bd-2166a8d093cb-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "ad155aa0-50a3-4874-93bd-2166a8d093cb" (UID: "ad155aa0-50a3-4874-93bd-2166a8d093cb"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.376852 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ad155aa0-50a3-4874-93bd-2166a8d093cb-oauth-serving-cert\") pod \"ad155aa0-50a3-4874-93bd-2166a8d093cb\" (UID: \"ad155aa0-50a3-4874-93bd-2166a8d093cb\") " Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.377003 4934 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad155aa0-50a3-4874-93bd-2166a8d093cb-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.377020 4934 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ad155aa0-50a3-4874-93bd-2166a8d093cb-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.377029 4934 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ad155aa0-50a3-4874-93bd-2166a8d093cb-console-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.377354 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad155aa0-50a3-4874-93bd-2166a8d093cb-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "ad155aa0-50a3-4874-93bd-2166a8d093cb" (UID: "ad155aa0-50a3-4874-93bd-2166a8d093cb"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.380367 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad155aa0-50a3-4874-93bd-2166a8d093cb-kube-api-access-kr89h" (OuterVolumeSpecName: "kube-api-access-kr89h") pod "ad155aa0-50a3-4874-93bd-2166a8d093cb" (UID: "ad155aa0-50a3-4874-93bd-2166a8d093cb"). InnerVolumeSpecName "kube-api-access-kr89h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.381282 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad155aa0-50a3-4874-93bd-2166a8d093cb-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "ad155aa0-50a3-4874-93bd-2166a8d093cb" (UID: "ad155aa0-50a3-4874-93bd-2166a8d093cb"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.381737 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad155aa0-50a3-4874-93bd-2166a8d093cb-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "ad155aa0-50a3-4874-93bd-2166a8d093cb" (UID: "ad155aa0-50a3-4874-93bd-2166a8d093cb"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.439830 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.439915 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.477662 4934 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ad155aa0-50a3-4874-93bd-2166a8d093cb-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.477717 4934 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ad155aa0-50a3-4874-93bd-2166a8d093cb-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.477730 4934 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ad155aa0-50a3-4874-93bd-2166a8d093cb-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.477741 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kr89h\" (UniqueName: \"kubernetes.io/projected/ad155aa0-50a3-4874-93bd-2166a8d093cb-kube-api-access-kr89h\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.532774 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r"] Oct 02 10:01:08 crc kubenswrapper[4934]: E1002 10:01:08.533191 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad155aa0-50a3-4874-93bd-2166a8d093cb" containerName="console" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.533232 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad155aa0-50a3-4874-93bd-2166a8d093cb" containerName="console" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.533500 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad155aa0-50a3-4874-93bd-2166a8d093cb" containerName="console" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.534859 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.538185 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.549781 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r"] Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.680479 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0107ccfd-b36f-4206-8f09-ff422dbc6a80-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r\" (UID: \"0107ccfd-b36f-4206-8f09-ff422dbc6a80\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.680722 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0107ccfd-b36f-4206-8f09-ff422dbc6a80-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r\" (UID: \"0107ccfd-b36f-4206-8f09-ff422dbc6a80\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.680796 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj7b8\" (UniqueName: \"kubernetes.io/projected/0107ccfd-b36f-4206-8f09-ff422dbc6a80-kube-api-access-gj7b8\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r\" (UID: \"0107ccfd-b36f-4206-8f09-ff422dbc6a80\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.781688 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-f44r4_ad155aa0-50a3-4874-93bd-2166a8d093cb/console/0.log" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.781744 4934 generic.go:334] "Generic (PLEG): container finished" podID="ad155aa0-50a3-4874-93bd-2166a8d093cb" containerID="4a7f374112692e2fb8e8ed98087537f84a98f191b103da6bc4f4be2c2d11d5cc" exitCode=2 Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.781694 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0107ccfd-b36f-4206-8f09-ff422dbc6a80-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r\" (UID: \"0107ccfd-b36f-4206-8f09-ff422dbc6a80\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.781854 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-f44r4" event={"ID":"ad155aa0-50a3-4874-93bd-2166a8d093cb","Type":"ContainerDied","Data":"4a7f374112692e2fb8e8ed98087537f84a98f191b103da6bc4f4be2c2d11d5cc"} Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.781863 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-f44r4" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.781915 4934 scope.go:117] "RemoveContainer" containerID="4a7f374112692e2fb8e8ed98087537f84a98f191b103da6bc4f4be2c2d11d5cc" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.781928 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj7b8\" (UniqueName: \"kubernetes.io/projected/0107ccfd-b36f-4206-8f09-ff422dbc6a80-kube-api-access-gj7b8\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r\" (UID: \"0107ccfd-b36f-4206-8f09-ff422dbc6a80\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.781973 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0107ccfd-b36f-4206-8f09-ff422dbc6a80-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r\" (UID: \"0107ccfd-b36f-4206-8f09-ff422dbc6a80\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.781901 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-f44r4" event={"ID":"ad155aa0-50a3-4874-93bd-2166a8d093cb","Type":"ContainerDied","Data":"a626b08f6728d9fc8d3ce28ee4ced96065cc97c159063fb5beb5eafd357f65b9"} Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.783498 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0107ccfd-b36f-4206-8f09-ff422dbc6a80-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r\" (UID: \"0107ccfd-b36f-4206-8f09-ff422dbc6a80\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.783715 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0107ccfd-b36f-4206-8f09-ff422dbc6a80-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r\" (UID: \"0107ccfd-b36f-4206-8f09-ff422dbc6a80\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.794786 4934 scope.go:117] "RemoveContainer" containerID="4a7f374112692e2fb8e8ed98087537f84a98f191b103da6bc4f4be2c2d11d5cc" Oct 02 10:01:08 crc kubenswrapper[4934]: E1002 10:01:08.795272 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a7f374112692e2fb8e8ed98087537f84a98f191b103da6bc4f4be2c2d11d5cc\": container with ID starting with 4a7f374112692e2fb8e8ed98087537f84a98f191b103da6bc4f4be2c2d11d5cc not found: ID does not exist" containerID="4a7f374112692e2fb8e8ed98087537f84a98f191b103da6bc4f4be2c2d11d5cc" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.795316 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a7f374112692e2fb8e8ed98087537f84a98f191b103da6bc4f4be2c2d11d5cc"} err="failed to get container status \"4a7f374112692e2fb8e8ed98087537f84a98f191b103da6bc4f4be2c2d11d5cc\": rpc error: code = NotFound desc = could not find container \"4a7f374112692e2fb8e8ed98087537f84a98f191b103da6bc4f4be2c2d11d5cc\": container with ID starting with 4a7f374112692e2fb8e8ed98087537f84a98f191b103da6bc4f4be2c2d11d5cc not found: ID does not exist" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.802310 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj7b8\" (UniqueName: \"kubernetes.io/projected/0107ccfd-b36f-4206-8f09-ff422dbc6a80-kube-api-access-gj7b8\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r\" (UID: \"0107ccfd-b36f-4206-8f09-ff422dbc6a80\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.816082 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-f44r4"] Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.819276 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-f44r4"] Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.857134 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r" Oct 02 10:01:08 crc kubenswrapper[4934]: I1002 10:01:08.926263 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad155aa0-50a3-4874-93bd-2166a8d093cb" path="/var/lib/kubelet/pods/ad155aa0-50a3-4874-93bd-2166a8d093cb/volumes" Oct 02 10:01:09 crc kubenswrapper[4934]: I1002 10:01:09.199709 4934 patch_prober.go:28] interesting pod/console-f9d7485db-f44r4 container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.28:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 02 10:01:09 crc kubenswrapper[4934]: I1002 10:01:09.199773 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-f9d7485db-f44r4" podUID="ad155aa0-50a3-4874-93bd-2166a8d093cb" containerName="console" probeResult="failure" output="Get \"https://10.217.0.28:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 02 10:01:09 crc kubenswrapper[4934]: I1002 10:01:09.256777 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r"] Oct 02 10:01:09 crc kubenswrapper[4934]: W1002 10:01:09.266673 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0107ccfd_b36f_4206_8f09_ff422dbc6a80.slice/crio-cd519b2d690dd73bf1725b6903779932e21a3f3dd3c5905491d9d453bcc6978c WatchSource:0}: Error finding container cd519b2d690dd73bf1725b6903779932e21a3f3dd3c5905491d9d453bcc6978c: Status 404 returned error can't find the container with id cd519b2d690dd73bf1725b6903779932e21a3f3dd3c5905491d9d453bcc6978c Oct 02 10:01:09 crc kubenswrapper[4934]: I1002 10:01:09.791830 4934 generic.go:334] "Generic (PLEG): container finished" podID="0107ccfd-b36f-4206-8f09-ff422dbc6a80" containerID="0d7db17792ef94dd07212e37093d68aa1aa6560c96aa5764cb8c2f1e7578ae57" exitCode=0 Oct 02 10:01:09 crc kubenswrapper[4934]: I1002 10:01:09.791881 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r" event={"ID":"0107ccfd-b36f-4206-8f09-ff422dbc6a80","Type":"ContainerDied","Data":"0d7db17792ef94dd07212e37093d68aa1aa6560c96aa5764cb8c2f1e7578ae57"} Oct 02 10:01:09 crc kubenswrapper[4934]: I1002 10:01:09.792151 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r" event={"ID":"0107ccfd-b36f-4206-8f09-ff422dbc6a80","Type":"ContainerStarted","Data":"cd519b2d690dd73bf1725b6903779932e21a3f3dd3c5905491d9d453bcc6978c"} Oct 02 10:01:11 crc kubenswrapper[4934]: I1002 10:01:11.804908 4934 generic.go:334] "Generic (PLEG): container finished" podID="0107ccfd-b36f-4206-8f09-ff422dbc6a80" containerID="95b931a127772e1e733493de6a7bdefd381893306a781a6d4efe616094787453" exitCode=0 Oct 02 10:01:11 crc kubenswrapper[4934]: I1002 10:01:11.804972 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r" event={"ID":"0107ccfd-b36f-4206-8f09-ff422dbc6a80","Type":"ContainerDied","Data":"95b931a127772e1e733493de6a7bdefd381893306a781a6d4efe616094787453"} Oct 02 10:01:12 crc kubenswrapper[4934]: I1002 10:01:12.815127 4934 generic.go:334] "Generic (PLEG): container finished" podID="0107ccfd-b36f-4206-8f09-ff422dbc6a80" containerID="a5b06207700bc0ca95b76ed26664db9e50681191ef4547a39c7628a95d762487" exitCode=0 Oct 02 10:01:12 crc kubenswrapper[4934]: I1002 10:01:12.815198 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r" event={"ID":"0107ccfd-b36f-4206-8f09-ff422dbc6a80","Type":"ContainerDied","Data":"a5b06207700bc0ca95b76ed26664db9e50681191ef4547a39c7628a95d762487"} Oct 02 10:01:13 crc kubenswrapper[4934]: I1002 10:01:13.271845 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7r6l7"] Oct 02 10:01:13 crc kubenswrapper[4934]: I1002 10:01:13.273385 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7r6l7" Oct 02 10:01:13 crc kubenswrapper[4934]: I1002 10:01:13.290656 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7r6l7"] Oct 02 10:01:13 crc kubenswrapper[4934]: I1002 10:01:13.440572 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skjld\" (UniqueName: \"kubernetes.io/projected/95a32e49-25af-4cd7-b306-35a22dc68f2b-kube-api-access-skjld\") pod \"redhat-operators-7r6l7\" (UID: \"95a32e49-25af-4cd7-b306-35a22dc68f2b\") " pod="openshift-marketplace/redhat-operators-7r6l7" Oct 02 10:01:13 crc kubenswrapper[4934]: I1002 10:01:13.440771 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95a32e49-25af-4cd7-b306-35a22dc68f2b-utilities\") pod \"redhat-operators-7r6l7\" (UID: \"95a32e49-25af-4cd7-b306-35a22dc68f2b\") " pod="openshift-marketplace/redhat-operators-7r6l7" Oct 02 10:01:13 crc kubenswrapper[4934]: I1002 10:01:13.440931 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95a32e49-25af-4cd7-b306-35a22dc68f2b-catalog-content\") pod \"redhat-operators-7r6l7\" (UID: \"95a32e49-25af-4cd7-b306-35a22dc68f2b\") " pod="openshift-marketplace/redhat-operators-7r6l7" Oct 02 10:01:13 crc kubenswrapper[4934]: I1002 10:01:13.542139 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95a32e49-25af-4cd7-b306-35a22dc68f2b-catalog-content\") pod \"redhat-operators-7r6l7\" (UID: \"95a32e49-25af-4cd7-b306-35a22dc68f2b\") " pod="openshift-marketplace/redhat-operators-7r6l7" Oct 02 10:01:13 crc kubenswrapper[4934]: I1002 10:01:13.542185 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skjld\" (UniqueName: \"kubernetes.io/projected/95a32e49-25af-4cd7-b306-35a22dc68f2b-kube-api-access-skjld\") pod \"redhat-operators-7r6l7\" (UID: \"95a32e49-25af-4cd7-b306-35a22dc68f2b\") " pod="openshift-marketplace/redhat-operators-7r6l7" Oct 02 10:01:13 crc kubenswrapper[4934]: I1002 10:01:13.542236 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95a32e49-25af-4cd7-b306-35a22dc68f2b-utilities\") pod \"redhat-operators-7r6l7\" (UID: \"95a32e49-25af-4cd7-b306-35a22dc68f2b\") " pod="openshift-marketplace/redhat-operators-7r6l7" Oct 02 10:01:13 crc kubenswrapper[4934]: I1002 10:01:13.542722 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95a32e49-25af-4cd7-b306-35a22dc68f2b-catalog-content\") pod \"redhat-operators-7r6l7\" (UID: \"95a32e49-25af-4cd7-b306-35a22dc68f2b\") " pod="openshift-marketplace/redhat-operators-7r6l7" Oct 02 10:01:13 crc kubenswrapper[4934]: I1002 10:01:13.542727 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95a32e49-25af-4cd7-b306-35a22dc68f2b-utilities\") pod \"redhat-operators-7r6l7\" (UID: \"95a32e49-25af-4cd7-b306-35a22dc68f2b\") " pod="openshift-marketplace/redhat-operators-7r6l7" Oct 02 10:01:13 crc kubenswrapper[4934]: I1002 10:01:13.561990 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skjld\" (UniqueName: \"kubernetes.io/projected/95a32e49-25af-4cd7-b306-35a22dc68f2b-kube-api-access-skjld\") pod \"redhat-operators-7r6l7\" (UID: \"95a32e49-25af-4cd7-b306-35a22dc68f2b\") " pod="openshift-marketplace/redhat-operators-7r6l7" Oct 02 10:01:13 crc kubenswrapper[4934]: I1002 10:01:13.606826 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7r6l7" Oct 02 10:01:14 crc kubenswrapper[4934]: I1002 10:01:14.064758 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7r6l7"] Oct 02 10:01:14 crc kubenswrapper[4934]: I1002 10:01:14.162042 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r" Oct 02 10:01:14 crc kubenswrapper[4934]: I1002 10:01:14.362484 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gj7b8\" (UniqueName: \"kubernetes.io/projected/0107ccfd-b36f-4206-8f09-ff422dbc6a80-kube-api-access-gj7b8\") pod \"0107ccfd-b36f-4206-8f09-ff422dbc6a80\" (UID: \"0107ccfd-b36f-4206-8f09-ff422dbc6a80\") " Oct 02 10:01:14 crc kubenswrapper[4934]: I1002 10:01:14.362526 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0107ccfd-b36f-4206-8f09-ff422dbc6a80-bundle\") pod \"0107ccfd-b36f-4206-8f09-ff422dbc6a80\" (UID: \"0107ccfd-b36f-4206-8f09-ff422dbc6a80\") " Oct 02 10:01:14 crc kubenswrapper[4934]: I1002 10:01:14.362547 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0107ccfd-b36f-4206-8f09-ff422dbc6a80-util\") pod \"0107ccfd-b36f-4206-8f09-ff422dbc6a80\" (UID: \"0107ccfd-b36f-4206-8f09-ff422dbc6a80\") " Oct 02 10:01:14 crc kubenswrapper[4934]: I1002 10:01:14.363520 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0107ccfd-b36f-4206-8f09-ff422dbc6a80-bundle" (OuterVolumeSpecName: "bundle") pod "0107ccfd-b36f-4206-8f09-ff422dbc6a80" (UID: "0107ccfd-b36f-4206-8f09-ff422dbc6a80"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:01:14 crc kubenswrapper[4934]: I1002 10:01:14.370277 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0107ccfd-b36f-4206-8f09-ff422dbc6a80-kube-api-access-gj7b8" (OuterVolumeSpecName: "kube-api-access-gj7b8") pod "0107ccfd-b36f-4206-8f09-ff422dbc6a80" (UID: "0107ccfd-b36f-4206-8f09-ff422dbc6a80"). InnerVolumeSpecName "kube-api-access-gj7b8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:01:14 crc kubenswrapper[4934]: I1002 10:01:14.383776 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0107ccfd-b36f-4206-8f09-ff422dbc6a80-util" (OuterVolumeSpecName: "util") pod "0107ccfd-b36f-4206-8f09-ff422dbc6a80" (UID: "0107ccfd-b36f-4206-8f09-ff422dbc6a80"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:01:14 crc kubenswrapper[4934]: I1002 10:01:14.464482 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gj7b8\" (UniqueName: \"kubernetes.io/projected/0107ccfd-b36f-4206-8f09-ff422dbc6a80-kube-api-access-gj7b8\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:14 crc kubenswrapper[4934]: I1002 10:01:14.464524 4934 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0107ccfd-b36f-4206-8f09-ff422dbc6a80-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:14 crc kubenswrapper[4934]: I1002 10:01:14.464536 4934 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0107ccfd-b36f-4206-8f09-ff422dbc6a80-util\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:14 crc kubenswrapper[4934]: I1002 10:01:14.828537 4934 generic.go:334] "Generic (PLEG): container finished" podID="95a32e49-25af-4cd7-b306-35a22dc68f2b" containerID="8a1f1d20a153a0a49eaae4976e38ad7b38f7c3869be074eda78607a212720f92" exitCode=0 Oct 02 10:01:14 crc kubenswrapper[4934]: I1002 10:01:14.828657 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7r6l7" event={"ID":"95a32e49-25af-4cd7-b306-35a22dc68f2b","Type":"ContainerDied","Data":"8a1f1d20a153a0a49eaae4976e38ad7b38f7c3869be074eda78607a212720f92"} Oct 02 10:01:14 crc kubenswrapper[4934]: I1002 10:01:14.828725 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7r6l7" event={"ID":"95a32e49-25af-4cd7-b306-35a22dc68f2b","Type":"ContainerStarted","Data":"bea4d658cc4bcb51df6bd3f1e7448784dee70fef16cdae05d1283b7daf9ebb2f"} Oct 02 10:01:14 crc kubenswrapper[4934]: I1002 10:01:14.831771 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r" event={"ID":"0107ccfd-b36f-4206-8f09-ff422dbc6a80","Type":"ContainerDied","Data":"cd519b2d690dd73bf1725b6903779932e21a3f3dd3c5905491d9d453bcc6978c"} Oct 02 10:01:14 crc kubenswrapper[4934]: I1002 10:01:14.831829 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd519b2d690dd73bf1725b6903779932e21a3f3dd3c5905491d9d453bcc6978c" Oct 02 10:01:14 crc kubenswrapper[4934]: I1002 10:01:14.831860 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r" Oct 02 10:01:14 crc kubenswrapper[4934]: I1002 10:01:14.998789 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7qdvp" Oct 02 10:01:14 crc kubenswrapper[4934]: I1002 10:01:14.998833 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7qdvp" Oct 02 10:01:15 crc kubenswrapper[4934]: I1002 10:01:15.044919 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7qdvp" Oct 02 10:01:15 crc kubenswrapper[4934]: I1002 10:01:15.908698 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7qdvp" Oct 02 10:01:16 crc kubenswrapper[4934]: I1002 10:01:16.851791 4934 generic.go:334] "Generic (PLEG): container finished" podID="95a32e49-25af-4cd7-b306-35a22dc68f2b" containerID="2c275284e2b3b992a453f78284b43f9eb7298a79467eafef5352bc25ff54fe12" exitCode=0 Oct 02 10:01:16 crc kubenswrapper[4934]: I1002 10:01:16.851953 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7r6l7" event={"ID":"95a32e49-25af-4cd7-b306-35a22dc68f2b","Type":"ContainerDied","Data":"2c275284e2b3b992a453f78284b43f9eb7298a79467eafef5352bc25ff54fe12"} Oct 02 10:01:18 crc kubenswrapper[4934]: I1002 10:01:18.901055 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7r6l7" event={"ID":"95a32e49-25af-4cd7-b306-35a22dc68f2b","Type":"ContainerStarted","Data":"1b5f1a9f03a02be791d1a1abc20671c41f407c53a83b06c3cb5415c8387b5a50"} Oct 02 10:01:18 crc kubenswrapper[4934]: I1002 10:01:18.931726 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7r6l7" podStartSLOduration=2.921294776 podStartE2EDuration="5.931708015s" podCreationTimestamp="2025-10-02 10:01:13 +0000 UTC" firstStartedPulling="2025-10-02 10:01:14.830678962 +0000 UTC m=+746.583320484" lastFinishedPulling="2025-10-02 10:01:17.841092211 +0000 UTC m=+749.593733723" observedRunningTime="2025-10-02 10:01:18.927081068 +0000 UTC m=+750.679722610" watchObservedRunningTime="2025-10-02 10:01:18.931708015 +0000 UTC m=+750.684349537" Oct 02 10:01:19 crc kubenswrapper[4934]: I1002 10:01:19.868312 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7qdvp"] Oct 02 10:01:19 crc kubenswrapper[4934]: I1002 10:01:19.868920 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7qdvp" podUID="c1888f99-8262-4f0e-97b8-fabef87b7557" containerName="registry-server" containerID="cri-o://d44f40b0a05d6078cca95175e8a987d65d9c7ebf9cf3db408d4a1f58c86988b5" gracePeriod=2 Oct 02 10:01:20 crc kubenswrapper[4934]: I1002 10:01:20.403869 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7qdvp" Oct 02 10:01:20 crc kubenswrapper[4934]: I1002 10:01:20.546040 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgbm9\" (UniqueName: \"kubernetes.io/projected/c1888f99-8262-4f0e-97b8-fabef87b7557-kube-api-access-tgbm9\") pod \"c1888f99-8262-4f0e-97b8-fabef87b7557\" (UID: \"c1888f99-8262-4f0e-97b8-fabef87b7557\") " Oct 02 10:01:20 crc kubenswrapper[4934]: I1002 10:01:20.546133 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1888f99-8262-4f0e-97b8-fabef87b7557-utilities\") pod \"c1888f99-8262-4f0e-97b8-fabef87b7557\" (UID: \"c1888f99-8262-4f0e-97b8-fabef87b7557\") " Oct 02 10:01:20 crc kubenswrapper[4934]: I1002 10:01:20.546161 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1888f99-8262-4f0e-97b8-fabef87b7557-catalog-content\") pod \"c1888f99-8262-4f0e-97b8-fabef87b7557\" (UID: \"c1888f99-8262-4f0e-97b8-fabef87b7557\") " Oct 02 10:01:20 crc kubenswrapper[4934]: I1002 10:01:20.547962 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1888f99-8262-4f0e-97b8-fabef87b7557-utilities" (OuterVolumeSpecName: "utilities") pod "c1888f99-8262-4f0e-97b8-fabef87b7557" (UID: "c1888f99-8262-4f0e-97b8-fabef87b7557"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:01:20 crc kubenswrapper[4934]: I1002 10:01:20.558742 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1888f99-8262-4f0e-97b8-fabef87b7557-kube-api-access-tgbm9" (OuterVolumeSpecName: "kube-api-access-tgbm9") pod "c1888f99-8262-4f0e-97b8-fabef87b7557" (UID: "c1888f99-8262-4f0e-97b8-fabef87b7557"). InnerVolumeSpecName "kube-api-access-tgbm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:01:20 crc kubenswrapper[4934]: I1002 10:01:20.609473 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1888f99-8262-4f0e-97b8-fabef87b7557-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c1888f99-8262-4f0e-97b8-fabef87b7557" (UID: "c1888f99-8262-4f0e-97b8-fabef87b7557"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:01:20 crc kubenswrapper[4934]: I1002 10:01:20.648166 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgbm9\" (UniqueName: \"kubernetes.io/projected/c1888f99-8262-4f0e-97b8-fabef87b7557-kube-api-access-tgbm9\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:20 crc kubenswrapper[4934]: I1002 10:01:20.648197 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1888f99-8262-4f0e-97b8-fabef87b7557-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:20 crc kubenswrapper[4934]: I1002 10:01:20.648206 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1888f99-8262-4f0e-97b8-fabef87b7557-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:20 crc kubenswrapper[4934]: I1002 10:01:20.914096 4934 generic.go:334] "Generic (PLEG): container finished" podID="c1888f99-8262-4f0e-97b8-fabef87b7557" containerID="d44f40b0a05d6078cca95175e8a987d65d9c7ebf9cf3db408d4a1f58c86988b5" exitCode=0 Oct 02 10:01:20 crc kubenswrapper[4934]: I1002 10:01:20.914226 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7qdvp" Oct 02 10:01:20 crc kubenswrapper[4934]: I1002 10:01:20.920138 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7qdvp" event={"ID":"c1888f99-8262-4f0e-97b8-fabef87b7557","Type":"ContainerDied","Data":"d44f40b0a05d6078cca95175e8a987d65d9c7ebf9cf3db408d4a1f58c86988b5"} Oct 02 10:01:20 crc kubenswrapper[4934]: I1002 10:01:20.920191 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7qdvp" event={"ID":"c1888f99-8262-4f0e-97b8-fabef87b7557","Type":"ContainerDied","Data":"a54efa35640922cecc7d46059ae35a2b478d75fb4b6191bb56e7aaade89e268f"} Oct 02 10:01:20 crc kubenswrapper[4934]: I1002 10:01:20.920216 4934 scope.go:117] "RemoveContainer" containerID="d44f40b0a05d6078cca95175e8a987d65d9c7ebf9cf3db408d4a1f58c86988b5" Oct 02 10:01:20 crc kubenswrapper[4934]: I1002 10:01:20.942872 4934 scope.go:117] "RemoveContainer" containerID="d1bfd10deb3932845a69155577eef144e9d3c8d6629ed7427562ff7e81779e4e" Oct 02 10:01:20 crc kubenswrapper[4934]: I1002 10:01:20.946538 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7qdvp"] Oct 02 10:01:20 crc kubenswrapper[4934]: I1002 10:01:20.952891 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7qdvp"] Oct 02 10:01:20 crc kubenswrapper[4934]: I1002 10:01:20.970098 4934 scope.go:117] "RemoveContainer" containerID="b60fc3bdbc4000acc6f1a506cb9d6e781c7dd640c8e449f619a13bb77d68164f" Oct 02 10:01:20 crc kubenswrapper[4934]: I1002 10:01:20.998762 4934 scope.go:117] "RemoveContainer" containerID="d44f40b0a05d6078cca95175e8a987d65d9c7ebf9cf3db408d4a1f58c86988b5" Oct 02 10:01:20 crc kubenswrapper[4934]: E1002 10:01:20.999316 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d44f40b0a05d6078cca95175e8a987d65d9c7ebf9cf3db408d4a1f58c86988b5\": container with ID starting with d44f40b0a05d6078cca95175e8a987d65d9c7ebf9cf3db408d4a1f58c86988b5 not found: ID does not exist" containerID="d44f40b0a05d6078cca95175e8a987d65d9c7ebf9cf3db408d4a1f58c86988b5" Oct 02 10:01:20 crc kubenswrapper[4934]: I1002 10:01:20.999357 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d44f40b0a05d6078cca95175e8a987d65d9c7ebf9cf3db408d4a1f58c86988b5"} err="failed to get container status \"d44f40b0a05d6078cca95175e8a987d65d9c7ebf9cf3db408d4a1f58c86988b5\": rpc error: code = NotFound desc = could not find container \"d44f40b0a05d6078cca95175e8a987d65d9c7ebf9cf3db408d4a1f58c86988b5\": container with ID starting with d44f40b0a05d6078cca95175e8a987d65d9c7ebf9cf3db408d4a1f58c86988b5 not found: ID does not exist" Oct 02 10:01:20 crc kubenswrapper[4934]: I1002 10:01:20.999397 4934 scope.go:117] "RemoveContainer" containerID="d1bfd10deb3932845a69155577eef144e9d3c8d6629ed7427562ff7e81779e4e" Oct 02 10:01:20 crc kubenswrapper[4934]: E1002 10:01:20.999840 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1bfd10deb3932845a69155577eef144e9d3c8d6629ed7427562ff7e81779e4e\": container with ID starting with d1bfd10deb3932845a69155577eef144e9d3c8d6629ed7427562ff7e81779e4e not found: ID does not exist" containerID="d1bfd10deb3932845a69155577eef144e9d3c8d6629ed7427562ff7e81779e4e" Oct 02 10:01:20 crc kubenswrapper[4934]: I1002 10:01:20.999876 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1bfd10deb3932845a69155577eef144e9d3c8d6629ed7427562ff7e81779e4e"} err="failed to get container status \"d1bfd10deb3932845a69155577eef144e9d3c8d6629ed7427562ff7e81779e4e\": rpc error: code = NotFound desc = could not find container \"d1bfd10deb3932845a69155577eef144e9d3c8d6629ed7427562ff7e81779e4e\": container with ID starting with d1bfd10deb3932845a69155577eef144e9d3c8d6629ed7427562ff7e81779e4e not found: ID does not exist" Oct 02 10:01:20 crc kubenswrapper[4934]: I1002 10:01:20.999910 4934 scope.go:117] "RemoveContainer" containerID="b60fc3bdbc4000acc6f1a506cb9d6e781c7dd640c8e449f619a13bb77d68164f" Oct 02 10:01:21 crc kubenswrapper[4934]: E1002 10:01:21.000265 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b60fc3bdbc4000acc6f1a506cb9d6e781c7dd640c8e449f619a13bb77d68164f\": container with ID starting with b60fc3bdbc4000acc6f1a506cb9d6e781c7dd640c8e449f619a13bb77d68164f not found: ID does not exist" containerID="b60fc3bdbc4000acc6f1a506cb9d6e781c7dd640c8e449f619a13bb77d68164f" Oct 02 10:01:21 crc kubenswrapper[4934]: I1002 10:01:21.000312 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b60fc3bdbc4000acc6f1a506cb9d6e781c7dd640c8e449f619a13bb77d68164f"} err="failed to get container status \"b60fc3bdbc4000acc6f1a506cb9d6e781c7dd640c8e449f619a13bb77d68164f\": rpc error: code = NotFound desc = could not find container \"b60fc3bdbc4000acc6f1a506cb9d6e781c7dd640c8e449f619a13bb77d68164f\": container with ID starting with b60fc3bdbc4000acc6f1a506cb9d6e781c7dd640c8e449f619a13bb77d68164f not found: ID does not exist" Oct 02 10:01:22 crc kubenswrapper[4934]: I1002 10:01:22.925875 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1888f99-8262-4f0e-97b8-fabef87b7557" path="/var/lib/kubelet/pods/c1888f99-8262-4f0e-97b8-fabef87b7557/volumes" Oct 02 10:01:23 crc kubenswrapper[4934]: I1002 10:01:23.608079 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7r6l7" Oct 02 10:01:23 crc kubenswrapper[4934]: I1002 10:01:23.608133 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7r6l7" Oct 02 10:01:23 crc kubenswrapper[4934]: I1002 10:01:23.661002 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7r6l7" Oct 02 10:01:23 crc kubenswrapper[4934]: I1002 10:01:23.963810 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7r6l7" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.297661 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-66d7dc4568-7b9dn"] Oct 02 10:01:24 crc kubenswrapper[4934]: E1002 10:01:24.297862 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0107ccfd-b36f-4206-8f09-ff422dbc6a80" containerName="pull" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.297873 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="0107ccfd-b36f-4206-8f09-ff422dbc6a80" containerName="pull" Oct 02 10:01:24 crc kubenswrapper[4934]: E1002 10:01:24.297884 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0107ccfd-b36f-4206-8f09-ff422dbc6a80" containerName="extract" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.297890 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="0107ccfd-b36f-4206-8f09-ff422dbc6a80" containerName="extract" Oct 02 10:01:24 crc kubenswrapper[4934]: E1002 10:01:24.297900 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1888f99-8262-4f0e-97b8-fabef87b7557" containerName="extract-content" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.297906 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1888f99-8262-4f0e-97b8-fabef87b7557" containerName="extract-content" Oct 02 10:01:24 crc kubenswrapper[4934]: E1002 10:01:24.297914 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1888f99-8262-4f0e-97b8-fabef87b7557" containerName="registry-server" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.297919 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1888f99-8262-4f0e-97b8-fabef87b7557" containerName="registry-server" Oct 02 10:01:24 crc kubenswrapper[4934]: E1002 10:01:24.297929 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0107ccfd-b36f-4206-8f09-ff422dbc6a80" containerName="util" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.297934 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="0107ccfd-b36f-4206-8f09-ff422dbc6a80" containerName="util" Oct 02 10:01:24 crc kubenswrapper[4934]: E1002 10:01:24.297956 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1888f99-8262-4f0e-97b8-fabef87b7557" containerName="extract-utilities" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.297962 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1888f99-8262-4f0e-97b8-fabef87b7557" containerName="extract-utilities" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.298046 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1888f99-8262-4f0e-97b8-fabef87b7557" containerName="registry-server" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.298056 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="0107ccfd-b36f-4206-8f09-ff422dbc6a80" containerName="extract" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.298422 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-66d7dc4568-7b9dn" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.301261 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.303554 4934 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.303890 4934 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.304296 4934 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-pd587" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.304753 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.320459 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-66d7dc4568-7b9dn"] Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.392070 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7fcd4706-2a88-4eda-9423-f535f059cbb3-webhook-cert\") pod \"metallb-operator-controller-manager-66d7dc4568-7b9dn\" (UID: \"7fcd4706-2a88-4eda-9423-f535f059cbb3\") " pod="metallb-system/metallb-operator-controller-manager-66d7dc4568-7b9dn" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.392325 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fskpn\" (UniqueName: \"kubernetes.io/projected/7fcd4706-2a88-4eda-9423-f535f059cbb3-kube-api-access-fskpn\") pod \"metallb-operator-controller-manager-66d7dc4568-7b9dn\" (UID: \"7fcd4706-2a88-4eda-9423-f535f059cbb3\") " pod="metallb-system/metallb-operator-controller-manager-66d7dc4568-7b9dn" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.392405 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7fcd4706-2a88-4eda-9423-f535f059cbb3-apiservice-cert\") pod \"metallb-operator-controller-manager-66d7dc4568-7b9dn\" (UID: \"7fcd4706-2a88-4eda-9423-f535f059cbb3\") " pod="metallb-system/metallb-operator-controller-manager-66d7dc4568-7b9dn" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.493475 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7fcd4706-2a88-4eda-9423-f535f059cbb3-webhook-cert\") pod \"metallb-operator-controller-manager-66d7dc4568-7b9dn\" (UID: \"7fcd4706-2a88-4eda-9423-f535f059cbb3\") " pod="metallb-system/metallb-operator-controller-manager-66d7dc4568-7b9dn" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.493524 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fskpn\" (UniqueName: \"kubernetes.io/projected/7fcd4706-2a88-4eda-9423-f535f059cbb3-kube-api-access-fskpn\") pod \"metallb-operator-controller-manager-66d7dc4568-7b9dn\" (UID: \"7fcd4706-2a88-4eda-9423-f535f059cbb3\") " pod="metallb-system/metallb-operator-controller-manager-66d7dc4568-7b9dn" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.493549 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7fcd4706-2a88-4eda-9423-f535f059cbb3-apiservice-cert\") pod \"metallb-operator-controller-manager-66d7dc4568-7b9dn\" (UID: \"7fcd4706-2a88-4eda-9423-f535f059cbb3\") " pod="metallb-system/metallb-operator-controller-manager-66d7dc4568-7b9dn" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.500874 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7fcd4706-2a88-4eda-9423-f535f059cbb3-apiservice-cert\") pod \"metallb-operator-controller-manager-66d7dc4568-7b9dn\" (UID: \"7fcd4706-2a88-4eda-9423-f535f059cbb3\") " pod="metallb-system/metallb-operator-controller-manager-66d7dc4568-7b9dn" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.512147 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fskpn\" (UniqueName: \"kubernetes.io/projected/7fcd4706-2a88-4eda-9423-f535f059cbb3-kube-api-access-fskpn\") pod \"metallb-operator-controller-manager-66d7dc4568-7b9dn\" (UID: \"7fcd4706-2a88-4eda-9423-f535f059cbb3\") " pod="metallb-system/metallb-operator-controller-manager-66d7dc4568-7b9dn" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.514523 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7fcd4706-2a88-4eda-9423-f535f059cbb3-webhook-cert\") pod \"metallb-operator-controller-manager-66d7dc4568-7b9dn\" (UID: \"7fcd4706-2a88-4eda-9423-f535f059cbb3\") " pod="metallb-system/metallb-operator-controller-manager-66d7dc4568-7b9dn" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.618519 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-66d7dc4568-7b9dn" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.640542 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-7b6d95f78d-q4xjj"] Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.641234 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7b6d95f78d-q4xjj" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.642670 4934 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-hvpk7" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.643754 4934 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.646829 4934 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.658914 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7b6d95f78d-q4xjj"] Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.798237 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b49e207c-8957-4b60-84f3-8b0b6c42579b-webhook-cert\") pod \"metallb-operator-webhook-server-7b6d95f78d-q4xjj\" (UID: \"b49e207c-8957-4b60-84f3-8b0b6c42579b\") " pod="metallb-system/metallb-operator-webhook-server-7b6d95f78d-q4xjj" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.798640 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc97c\" (UniqueName: \"kubernetes.io/projected/b49e207c-8957-4b60-84f3-8b0b6c42579b-kube-api-access-lc97c\") pod \"metallb-operator-webhook-server-7b6d95f78d-q4xjj\" (UID: \"b49e207c-8957-4b60-84f3-8b0b6c42579b\") " pod="metallb-system/metallb-operator-webhook-server-7b6d95f78d-q4xjj" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.798752 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b49e207c-8957-4b60-84f3-8b0b6c42579b-apiservice-cert\") pod \"metallb-operator-webhook-server-7b6d95f78d-q4xjj\" (UID: \"b49e207c-8957-4b60-84f3-8b0b6c42579b\") " pod="metallb-system/metallb-operator-webhook-server-7b6d95f78d-q4xjj" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.900083 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b49e207c-8957-4b60-84f3-8b0b6c42579b-webhook-cert\") pod \"metallb-operator-webhook-server-7b6d95f78d-q4xjj\" (UID: \"b49e207c-8957-4b60-84f3-8b0b6c42579b\") " pod="metallb-system/metallb-operator-webhook-server-7b6d95f78d-q4xjj" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.900191 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc97c\" (UniqueName: \"kubernetes.io/projected/b49e207c-8957-4b60-84f3-8b0b6c42579b-kube-api-access-lc97c\") pod \"metallb-operator-webhook-server-7b6d95f78d-q4xjj\" (UID: \"b49e207c-8957-4b60-84f3-8b0b6c42579b\") " pod="metallb-system/metallb-operator-webhook-server-7b6d95f78d-q4xjj" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.900380 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b49e207c-8957-4b60-84f3-8b0b6c42579b-apiservice-cert\") pod \"metallb-operator-webhook-server-7b6d95f78d-q4xjj\" (UID: \"b49e207c-8957-4b60-84f3-8b0b6c42579b\") " pod="metallb-system/metallb-operator-webhook-server-7b6d95f78d-q4xjj" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.904876 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b49e207c-8957-4b60-84f3-8b0b6c42579b-apiservice-cert\") pod \"metallb-operator-webhook-server-7b6d95f78d-q4xjj\" (UID: \"b49e207c-8957-4b60-84f3-8b0b6c42579b\") " pod="metallb-system/metallb-operator-webhook-server-7b6d95f78d-q4xjj" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.906167 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b49e207c-8957-4b60-84f3-8b0b6c42579b-webhook-cert\") pod \"metallb-operator-webhook-server-7b6d95f78d-q4xjj\" (UID: \"b49e207c-8957-4b60-84f3-8b0b6c42579b\") " pod="metallb-system/metallb-operator-webhook-server-7b6d95f78d-q4xjj" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.925910 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc97c\" (UniqueName: \"kubernetes.io/projected/b49e207c-8957-4b60-84f3-8b0b6c42579b-kube-api-access-lc97c\") pod \"metallb-operator-webhook-server-7b6d95f78d-q4xjj\" (UID: \"b49e207c-8957-4b60-84f3-8b0b6c42579b\") " pod="metallb-system/metallb-operator-webhook-server-7b6d95f78d-q4xjj" Oct 02 10:01:24 crc kubenswrapper[4934]: I1002 10:01:24.986439 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7b6d95f78d-q4xjj" Oct 02 10:01:25 crc kubenswrapper[4934]: I1002 10:01:25.077960 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-66d7dc4568-7b9dn"] Oct 02 10:01:25 crc kubenswrapper[4934]: W1002 10:01:25.086439 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fcd4706_2a88_4eda_9423_f535f059cbb3.slice/crio-5b8571f634c1218305f1b51e61f45066058115b1aade88b3cae999101d4b21ea WatchSource:0}: Error finding container 5b8571f634c1218305f1b51e61f45066058115b1aade88b3cae999101d4b21ea: Status 404 returned error can't find the container with id 5b8571f634c1218305f1b51e61f45066058115b1aade88b3cae999101d4b21ea Oct 02 10:01:25 crc kubenswrapper[4934]: I1002 10:01:25.400798 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7b6d95f78d-q4xjj"] Oct 02 10:01:25 crc kubenswrapper[4934]: W1002 10:01:25.405288 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb49e207c_8957_4b60_84f3_8b0b6c42579b.slice/crio-62748ead88899bf248c96c6b364ebd65ba30f2f7d86162267da1a585cd93a520 WatchSource:0}: Error finding container 62748ead88899bf248c96c6b364ebd65ba30f2f7d86162267da1a585cd93a520: Status 404 returned error can't find the container with id 62748ead88899bf248c96c6b364ebd65ba30f2f7d86162267da1a585cd93a520 Oct 02 10:01:25 crc kubenswrapper[4934]: I1002 10:01:25.460285 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7r6l7"] Oct 02 10:01:25 crc kubenswrapper[4934]: I1002 10:01:25.940816 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7b6d95f78d-q4xjj" event={"ID":"b49e207c-8957-4b60-84f3-8b0b6c42579b","Type":"ContainerStarted","Data":"62748ead88899bf248c96c6b364ebd65ba30f2f7d86162267da1a585cd93a520"} Oct 02 10:01:25 crc kubenswrapper[4934]: I1002 10:01:25.942002 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-66d7dc4568-7b9dn" event={"ID":"7fcd4706-2a88-4eda-9423-f535f059cbb3","Type":"ContainerStarted","Data":"5b8571f634c1218305f1b51e61f45066058115b1aade88b3cae999101d4b21ea"} Oct 02 10:01:25 crc kubenswrapper[4934]: I1002 10:01:25.942247 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7r6l7" podUID="95a32e49-25af-4cd7-b306-35a22dc68f2b" containerName="registry-server" containerID="cri-o://1b5f1a9f03a02be791d1a1abc20671c41f407c53a83b06c3cb5415c8387b5a50" gracePeriod=2 Oct 02 10:01:26 crc kubenswrapper[4934]: I1002 10:01:26.443553 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7r6l7" Oct 02 10:01:26 crc kubenswrapper[4934]: I1002 10:01:26.625964 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skjld\" (UniqueName: \"kubernetes.io/projected/95a32e49-25af-4cd7-b306-35a22dc68f2b-kube-api-access-skjld\") pod \"95a32e49-25af-4cd7-b306-35a22dc68f2b\" (UID: \"95a32e49-25af-4cd7-b306-35a22dc68f2b\") " Oct 02 10:01:26 crc kubenswrapper[4934]: I1002 10:01:26.626037 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95a32e49-25af-4cd7-b306-35a22dc68f2b-utilities\") pod \"95a32e49-25af-4cd7-b306-35a22dc68f2b\" (UID: \"95a32e49-25af-4cd7-b306-35a22dc68f2b\") " Oct 02 10:01:26 crc kubenswrapper[4934]: I1002 10:01:26.626085 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95a32e49-25af-4cd7-b306-35a22dc68f2b-catalog-content\") pod \"95a32e49-25af-4cd7-b306-35a22dc68f2b\" (UID: \"95a32e49-25af-4cd7-b306-35a22dc68f2b\") " Oct 02 10:01:26 crc kubenswrapper[4934]: I1002 10:01:26.629924 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95a32e49-25af-4cd7-b306-35a22dc68f2b-utilities" (OuterVolumeSpecName: "utilities") pod "95a32e49-25af-4cd7-b306-35a22dc68f2b" (UID: "95a32e49-25af-4cd7-b306-35a22dc68f2b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:01:26 crc kubenswrapper[4934]: I1002 10:01:26.633146 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95a32e49-25af-4cd7-b306-35a22dc68f2b-kube-api-access-skjld" (OuterVolumeSpecName: "kube-api-access-skjld") pod "95a32e49-25af-4cd7-b306-35a22dc68f2b" (UID: "95a32e49-25af-4cd7-b306-35a22dc68f2b"). InnerVolumeSpecName "kube-api-access-skjld". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:01:26 crc kubenswrapper[4934]: I1002 10:01:26.730241 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skjld\" (UniqueName: \"kubernetes.io/projected/95a32e49-25af-4cd7-b306-35a22dc68f2b-kube-api-access-skjld\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:26 crc kubenswrapper[4934]: I1002 10:01:26.730594 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95a32e49-25af-4cd7-b306-35a22dc68f2b-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:26 crc kubenswrapper[4934]: I1002 10:01:26.951168 4934 generic.go:334] "Generic (PLEG): container finished" podID="95a32e49-25af-4cd7-b306-35a22dc68f2b" containerID="1b5f1a9f03a02be791d1a1abc20671c41f407c53a83b06c3cb5415c8387b5a50" exitCode=0 Oct 02 10:01:26 crc kubenswrapper[4934]: I1002 10:01:26.951214 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7r6l7" event={"ID":"95a32e49-25af-4cd7-b306-35a22dc68f2b","Type":"ContainerDied","Data":"1b5f1a9f03a02be791d1a1abc20671c41f407c53a83b06c3cb5415c8387b5a50"} Oct 02 10:01:26 crc kubenswrapper[4934]: I1002 10:01:26.951243 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7r6l7" event={"ID":"95a32e49-25af-4cd7-b306-35a22dc68f2b","Type":"ContainerDied","Data":"bea4d658cc4bcb51df6bd3f1e7448784dee70fef16cdae05d1283b7daf9ebb2f"} Oct 02 10:01:26 crc kubenswrapper[4934]: I1002 10:01:26.951262 4934 scope.go:117] "RemoveContainer" containerID="1b5f1a9f03a02be791d1a1abc20671c41f407c53a83b06c3cb5415c8387b5a50" Oct 02 10:01:26 crc kubenswrapper[4934]: I1002 10:01:26.951378 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7r6l7" Oct 02 10:01:26 crc kubenswrapper[4934]: I1002 10:01:26.973115 4934 scope.go:117] "RemoveContainer" containerID="2c275284e2b3b992a453f78284b43f9eb7298a79467eafef5352bc25ff54fe12" Oct 02 10:01:26 crc kubenswrapper[4934]: I1002 10:01:26.989442 4934 scope.go:117] "RemoveContainer" containerID="8a1f1d20a153a0a49eaae4976e38ad7b38f7c3869be074eda78607a212720f92" Oct 02 10:01:27 crc kubenswrapper[4934]: I1002 10:01:27.029933 4934 scope.go:117] "RemoveContainer" containerID="1b5f1a9f03a02be791d1a1abc20671c41f407c53a83b06c3cb5415c8387b5a50" Oct 02 10:01:27 crc kubenswrapper[4934]: E1002 10:01:27.030526 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b5f1a9f03a02be791d1a1abc20671c41f407c53a83b06c3cb5415c8387b5a50\": container with ID starting with 1b5f1a9f03a02be791d1a1abc20671c41f407c53a83b06c3cb5415c8387b5a50 not found: ID does not exist" containerID="1b5f1a9f03a02be791d1a1abc20671c41f407c53a83b06c3cb5415c8387b5a50" Oct 02 10:01:27 crc kubenswrapper[4934]: I1002 10:01:27.030691 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b5f1a9f03a02be791d1a1abc20671c41f407c53a83b06c3cb5415c8387b5a50"} err="failed to get container status \"1b5f1a9f03a02be791d1a1abc20671c41f407c53a83b06c3cb5415c8387b5a50\": rpc error: code = NotFound desc = could not find container \"1b5f1a9f03a02be791d1a1abc20671c41f407c53a83b06c3cb5415c8387b5a50\": container with ID starting with 1b5f1a9f03a02be791d1a1abc20671c41f407c53a83b06c3cb5415c8387b5a50 not found: ID does not exist" Oct 02 10:01:27 crc kubenswrapper[4934]: I1002 10:01:27.030724 4934 scope.go:117] "RemoveContainer" containerID="2c275284e2b3b992a453f78284b43f9eb7298a79467eafef5352bc25ff54fe12" Oct 02 10:01:27 crc kubenswrapper[4934]: E1002 10:01:27.031044 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c275284e2b3b992a453f78284b43f9eb7298a79467eafef5352bc25ff54fe12\": container with ID starting with 2c275284e2b3b992a453f78284b43f9eb7298a79467eafef5352bc25ff54fe12 not found: ID does not exist" containerID="2c275284e2b3b992a453f78284b43f9eb7298a79467eafef5352bc25ff54fe12" Oct 02 10:01:27 crc kubenswrapper[4934]: I1002 10:01:27.031065 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c275284e2b3b992a453f78284b43f9eb7298a79467eafef5352bc25ff54fe12"} err="failed to get container status \"2c275284e2b3b992a453f78284b43f9eb7298a79467eafef5352bc25ff54fe12\": rpc error: code = NotFound desc = could not find container \"2c275284e2b3b992a453f78284b43f9eb7298a79467eafef5352bc25ff54fe12\": container with ID starting with 2c275284e2b3b992a453f78284b43f9eb7298a79467eafef5352bc25ff54fe12 not found: ID does not exist" Oct 02 10:01:27 crc kubenswrapper[4934]: I1002 10:01:27.031081 4934 scope.go:117] "RemoveContainer" containerID="8a1f1d20a153a0a49eaae4976e38ad7b38f7c3869be074eda78607a212720f92" Oct 02 10:01:27 crc kubenswrapper[4934]: E1002 10:01:27.031417 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a1f1d20a153a0a49eaae4976e38ad7b38f7c3869be074eda78607a212720f92\": container with ID starting with 8a1f1d20a153a0a49eaae4976e38ad7b38f7c3869be074eda78607a212720f92 not found: ID does not exist" containerID="8a1f1d20a153a0a49eaae4976e38ad7b38f7c3869be074eda78607a212720f92" Oct 02 10:01:27 crc kubenswrapper[4934]: I1002 10:01:27.031440 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a1f1d20a153a0a49eaae4976e38ad7b38f7c3869be074eda78607a212720f92"} err="failed to get container status \"8a1f1d20a153a0a49eaae4976e38ad7b38f7c3869be074eda78607a212720f92\": rpc error: code = NotFound desc = could not find container \"8a1f1d20a153a0a49eaae4976e38ad7b38f7c3869be074eda78607a212720f92\": container with ID starting with 8a1f1d20a153a0a49eaae4976e38ad7b38f7c3869be074eda78607a212720f92 not found: ID does not exist" Oct 02 10:01:28 crc kubenswrapper[4934]: I1002 10:01:28.605665 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95a32e49-25af-4cd7-b306-35a22dc68f2b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "95a32e49-25af-4cd7-b306-35a22dc68f2b" (UID: "95a32e49-25af-4cd7-b306-35a22dc68f2b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:01:28 crc kubenswrapper[4934]: I1002 10:01:28.665368 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95a32e49-25af-4cd7-b306-35a22dc68f2b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:28 crc kubenswrapper[4934]: I1002 10:01:28.782755 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7r6l7"] Oct 02 10:01:28 crc kubenswrapper[4934]: I1002 10:01:28.786030 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7r6l7"] Oct 02 10:01:28 crc kubenswrapper[4934]: I1002 10:01:28.920105 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95a32e49-25af-4cd7-b306-35a22dc68f2b" path="/var/lib/kubelet/pods/95a32e49-25af-4cd7-b306-35a22dc68f2b/volumes" Oct 02 10:01:30 crc kubenswrapper[4934]: I1002 10:01:30.974923 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7b6d95f78d-q4xjj" event={"ID":"b49e207c-8957-4b60-84f3-8b0b6c42579b","Type":"ContainerStarted","Data":"af5a8565fc4bbed8a87e39744aeb50151276853a9ddcc3a42a66f592c91161dd"} Oct 02 10:01:30 crc kubenswrapper[4934]: I1002 10:01:30.975175 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-7b6d95f78d-q4xjj" Oct 02 10:01:30 crc kubenswrapper[4934]: I1002 10:01:30.976460 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-66d7dc4568-7b9dn" event={"ID":"7fcd4706-2a88-4eda-9423-f535f059cbb3","Type":"ContainerStarted","Data":"78fb077ae9e78d62693c276e500b70b8173a3b0dabc65cfca89f3bbee97c1656"} Oct 02 10:01:30 crc kubenswrapper[4934]: I1002 10:01:30.976622 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-66d7dc4568-7b9dn" Oct 02 10:01:30 crc kubenswrapper[4934]: I1002 10:01:30.992280 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-7b6d95f78d-q4xjj" podStartSLOduration=2.341912439 podStartE2EDuration="6.99225675s" podCreationTimestamp="2025-10-02 10:01:24 +0000 UTC" firstStartedPulling="2025-10-02 10:01:25.409621318 +0000 UTC m=+757.162262840" lastFinishedPulling="2025-10-02 10:01:30.059965629 +0000 UTC m=+761.812607151" observedRunningTime="2025-10-02 10:01:30.991809548 +0000 UTC m=+762.744451070" watchObservedRunningTime="2025-10-02 10:01:30.99225675 +0000 UTC m=+762.744898282" Oct 02 10:01:31 crc kubenswrapper[4934]: I1002 10:01:31.018329 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-66d7dc4568-7b9dn" podStartSLOduration=2.047008802 podStartE2EDuration="7.018306718s" podCreationTimestamp="2025-10-02 10:01:24 +0000 UTC" firstStartedPulling="2025-10-02 10:01:25.090467873 +0000 UTC m=+756.843109395" lastFinishedPulling="2025-10-02 10:01:30.061765789 +0000 UTC m=+761.814407311" observedRunningTime="2025-10-02 10:01:31.016940451 +0000 UTC m=+762.769582003" watchObservedRunningTime="2025-10-02 10:01:31.018306718 +0000 UTC m=+762.770948260" Oct 02 10:01:38 crc kubenswrapper[4934]: I1002 10:01:38.439613 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:01:38 crc kubenswrapper[4934]: I1002 10:01:38.440411 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:01:38 crc kubenswrapper[4934]: I1002 10:01:38.440490 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 10:01:38 crc kubenswrapper[4934]: I1002 10:01:38.441332 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3834d18542216db82e6fdc0858cf6c8272b05593949e3231352e2c84300c22de"} pod="openshift-machine-config-operator/machine-config-daemon-djh5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 10:01:38 crc kubenswrapper[4934]: I1002 10:01:38.441424 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" containerID="cri-o://3834d18542216db82e6fdc0858cf6c8272b05593949e3231352e2c84300c22de" gracePeriod=600 Oct 02 10:01:39 crc kubenswrapper[4934]: I1002 10:01:39.020242 4934 generic.go:334] "Generic (PLEG): container finished" podID="71db06ef-05b0-4f58-b251-b27117a8500a" containerID="3834d18542216db82e6fdc0858cf6c8272b05593949e3231352e2c84300c22de" exitCode=0 Oct 02 10:01:39 crc kubenswrapper[4934]: I1002 10:01:39.020274 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerDied","Data":"3834d18542216db82e6fdc0858cf6c8272b05593949e3231352e2c84300c22de"} Oct 02 10:01:39 crc kubenswrapper[4934]: I1002 10:01:39.020546 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerStarted","Data":"5567160fec03506e3dac2e60d3a06fa8b4a72b51448a03102f357a2e1f66bb4b"} Oct 02 10:01:39 crc kubenswrapper[4934]: I1002 10:01:39.020567 4934 scope.go:117] "RemoveContainer" containerID="5d64ebbdf408852984cc94e96c4d8071b186211b51e5b9818eed5258b24a72f5" Oct 02 10:01:44 crc kubenswrapper[4934]: I1002 10:01:44.991562 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-7b6d95f78d-q4xjj" Oct 02 10:01:55 crc kubenswrapper[4934]: I1002 10:01:55.872486 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6hkdz"] Oct 02 10:01:55 crc kubenswrapper[4934]: E1002 10:01:55.874505 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95a32e49-25af-4cd7-b306-35a22dc68f2b" containerName="extract-utilities" Oct 02 10:01:55 crc kubenswrapper[4934]: I1002 10:01:55.874626 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="95a32e49-25af-4cd7-b306-35a22dc68f2b" containerName="extract-utilities" Oct 02 10:01:55 crc kubenswrapper[4934]: E1002 10:01:55.874714 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95a32e49-25af-4cd7-b306-35a22dc68f2b" containerName="registry-server" Oct 02 10:01:55 crc kubenswrapper[4934]: I1002 10:01:55.874785 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="95a32e49-25af-4cd7-b306-35a22dc68f2b" containerName="registry-server" Oct 02 10:01:55 crc kubenswrapper[4934]: E1002 10:01:55.874877 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95a32e49-25af-4cd7-b306-35a22dc68f2b" containerName="extract-content" Oct 02 10:01:55 crc kubenswrapper[4934]: I1002 10:01:55.874947 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="95a32e49-25af-4cd7-b306-35a22dc68f2b" containerName="extract-content" Oct 02 10:01:55 crc kubenswrapper[4934]: I1002 10:01:55.875160 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="95a32e49-25af-4cd7-b306-35a22dc68f2b" containerName="registry-server" Oct 02 10:01:55 crc kubenswrapper[4934]: I1002 10:01:55.876136 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6hkdz" Oct 02 10:01:55 crc kubenswrapper[4934]: I1002 10:01:55.893615 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6hkdz"] Oct 02 10:01:56 crc kubenswrapper[4934]: I1002 10:01:56.006528 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5be6c267-644a-4535-b143-0f70ed2e7efd-utilities\") pod \"redhat-marketplace-6hkdz\" (UID: \"5be6c267-644a-4535-b143-0f70ed2e7efd\") " pod="openshift-marketplace/redhat-marketplace-6hkdz" Oct 02 10:01:56 crc kubenswrapper[4934]: I1002 10:01:56.007227 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sg5g9\" (UniqueName: \"kubernetes.io/projected/5be6c267-644a-4535-b143-0f70ed2e7efd-kube-api-access-sg5g9\") pod \"redhat-marketplace-6hkdz\" (UID: \"5be6c267-644a-4535-b143-0f70ed2e7efd\") " pod="openshift-marketplace/redhat-marketplace-6hkdz" Oct 02 10:01:56 crc kubenswrapper[4934]: I1002 10:01:56.007484 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5be6c267-644a-4535-b143-0f70ed2e7efd-catalog-content\") pod \"redhat-marketplace-6hkdz\" (UID: \"5be6c267-644a-4535-b143-0f70ed2e7efd\") " pod="openshift-marketplace/redhat-marketplace-6hkdz" Oct 02 10:01:56 crc kubenswrapper[4934]: I1002 10:01:56.108857 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5be6c267-644a-4535-b143-0f70ed2e7efd-utilities\") pod \"redhat-marketplace-6hkdz\" (UID: \"5be6c267-644a-4535-b143-0f70ed2e7efd\") " pod="openshift-marketplace/redhat-marketplace-6hkdz" Oct 02 10:01:56 crc kubenswrapper[4934]: I1002 10:01:56.109174 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sg5g9\" (UniqueName: \"kubernetes.io/projected/5be6c267-644a-4535-b143-0f70ed2e7efd-kube-api-access-sg5g9\") pod \"redhat-marketplace-6hkdz\" (UID: \"5be6c267-644a-4535-b143-0f70ed2e7efd\") " pod="openshift-marketplace/redhat-marketplace-6hkdz" Oct 02 10:01:56 crc kubenswrapper[4934]: I1002 10:01:56.109320 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5be6c267-644a-4535-b143-0f70ed2e7efd-catalog-content\") pod \"redhat-marketplace-6hkdz\" (UID: \"5be6c267-644a-4535-b143-0f70ed2e7efd\") " pod="openshift-marketplace/redhat-marketplace-6hkdz" Oct 02 10:01:56 crc kubenswrapper[4934]: I1002 10:01:56.109463 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5be6c267-644a-4535-b143-0f70ed2e7efd-utilities\") pod \"redhat-marketplace-6hkdz\" (UID: \"5be6c267-644a-4535-b143-0f70ed2e7efd\") " pod="openshift-marketplace/redhat-marketplace-6hkdz" Oct 02 10:01:56 crc kubenswrapper[4934]: I1002 10:01:56.109739 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5be6c267-644a-4535-b143-0f70ed2e7efd-catalog-content\") pod \"redhat-marketplace-6hkdz\" (UID: \"5be6c267-644a-4535-b143-0f70ed2e7efd\") " pod="openshift-marketplace/redhat-marketplace-6hkdz" Oct 02 10:01:56 crc kubenswrapper[4934]: I1002 10:01:56.130877 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sg5g9\" (UniqueName: \"kubernetes.io/projected/5be6c267-644a-4535-b143-0f70ed2e7efd-kube-api-access-sg5g9\") pod \"redhat-marketplace-6hkdz\" (UID: \"5be6c267-644a-4535-b143-0f70ed2e7efd\") " pod="openshift-marketplace/redhat-marketplace-6hkdz" Oct 02 10:01:56 crc kubenswrapper[4934]: I1002 10:01:56.199323 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6hkdz" Oct 02 10:01:56 crc kubenswrapper[4934]: I1002 10:01:56.607118 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6hkdz"] Oct 02 10:01:57 crc kubenswrapper[4934]: I1002 10:01:57.130696 4934 generic.go:334] "Generic (PLEG): container finished" podID="5be6c267-644a-4535-b143-0f70ed2e7efd" containerID="605e2b7746b83422eaaa5eb57ebe09949a45d4da05f770f12d9bf09b01c3324f" exitCode=0 Oct 02 10:01:57 crc kubenswrapper[4934]: I1002 10:01:57.130761 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hkdz" event={"ID":"5be6c267-644a-4535-b143-0f70ed2e7efd","Type":"ContainerDied","Data":"605e2b7746b83422eaaa5eb57ebe09949a45d4da05f770f12d9bf09b01c3324f"} Oct 02 10:01:57 crc kubenswrapper[4934]: I1002 10:01:57.132554 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hkdz" event={"ID":"5be6c267-644a-4535-b143-0f70ed2e7efd","Type":"ContainerStarted","Data":"084a43750c18c053d28e7ef7d580d1425ae2344beb885ebbcaaa142296516ffc"} Oct 02 10:01:58 crc kubenswrapper[4934]: I1002 10:01:58.139895 4934 generic.go:334] "Generic (PLEG): container finished" podID="5be6c267-644a-4535-b143-0f70ed2e7efd" containerID="ed11ac14839ae582fa8a07660889121a912b2747f11f2323c636c47866b576db" exitCode=0 Oct 02 10:01:58 crc kubenswrapper[4934]: I1002 10:01:58.139984 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hkdz" event={"ID":"5be6c267-644a-4535-b143-0f70ed2e7efd","Type":"ContainerDied","Data":"ed11ac14839ae582fa8a07660889121a912b2747f11f2323c636c47866b576db"} Oct 02 10:01:59 crc kubenswrapper[4934]: I1002 10:01:59.147922 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hkdz" event={"ID":"5be6c267-644a-4535-b143-0f70ed2e7efd","Type":"ContainerStarted","Data":"5cee34364539c42a7dbce13d979203513de088356f120f9cd21e245eba1be240"} Oct 02 10:01:59 crc kubenswrapper[4934]: I1002 10:01:59.171410 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6hkdz" podStartSLOduration=2.755748778 podStartE2EDuration="4.171386939s" podCreationTimestamp="2025-10-02 10:01:55 +0000 UTC" firstStartedPulling="2025-10-02 10:01:57.132555158 +0000 UTC m=+788.885196690" lastFinishedPulling="2025-10-02 10:01:58.548193309 +0000 UTC m=+790.300834851" observedRunningTime="2025-10-02 10:01:59.170562855 +0000 UTC m=+790.923204467" watchObservedRunningTime="2025-10-02 10:01:59.171386939 +0000 UTC m=+790.924028491" Oct 02 10:02:04 crc kubenswrapper[4934]: I1002 10:02:04.621282 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-66d7dc4568-7b9dn" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.433319 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-pljrz"] Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.434141 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pljrz" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.437515 4934 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.437704 4934 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-kf6nr" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.438359 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-jlmg7"] Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.441343 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-jlmg7" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.443143 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-pljrz"] Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.443283 4934 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.443702 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.557599 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-m6ncq"] Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.558810 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-m6ncq" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.563425 4934 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.563467 4934 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.564059 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.568862 4934 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-kz46t" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.580370 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-74qhz"] Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.581258 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-74qhz" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.582939 4934 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.596286 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-74qhz"] Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.622504 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/853c1c50-69e1-47ca-9951-c75e52e11cc2-frr-sockets\") pod \"frr-k8s-jlmg7\" (UID: \"853c1c50-69e1-47ca-9951-c75e52e11cc2\") " pod="metallb-system/frr-k8s-jlmg7" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.622547 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/853c1c50-69e1-47ca-9951-c75e52e11cc2-frr-conf\") pod \"frr-k8s-jlmg7\" (UID: \"853c1c50-69e1-47ca-9951-c75e52e11cc2\") " pod="metallb-system/frr-k8s-jlmg7" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.622571 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/8600909e-f2a1-4472-89ca-275d753fc807-metallb-excludel2\") pod \"speaker-m6ncq\" (UID: \"8600909e-f2a1-4472-89ca-275d753fc807\") " pod="metallb-system/speaker-m6ncq" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.622657 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/853c1c50-69e1-47ca-9951-c75e52e11cc2-metrics-certs\") pod \"frr-k8s-jlmg7\" (UID: \"853c1c50-69e1-47ca-9951-c75e52e11cc2\") " pod="metallb-system/frr-k8s-jlmg7" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.622696 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8600909e-f2a1-4472-89ca-275d753fc807-metrics-certs\") pod \"speaker-m6ncq\" (UID: \"8600909e-f2a1-4472-89ca-275d753fc807\") " pod="metallb-system/speaker-m6ncq" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.622715 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/853c1c50-69e1-47ca-9951-c75e52e11cc2-metrics\") pod \"frr-k8s-jlmg7\" (UID: \"853c1c50-69e1-47ca-9951-c75e52e11cc2\") " pod="metallb-system/frr-k8s-jlmg7" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.622732 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7brt\" (UniqueName: \"kubernetes.io/projected/18786238-f18b-4b0a-b54d-e838c1092722-kube-api-access-d7brt\") pod \"controller-68d546b9d8-74qhz\" (UID: \"18786238-f18b-4b0a-b54d-e838c1092722\") " pod="metallb-system/controller-68d546b9d8-74qhz" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.622769 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/18786238-f18b-4b0a-b54d-e838c1092722-cert\") pod \"controller-68d546b9d8-74qhz\" (UID: \"18786238-f18b-4b0a-b54d-e838c1092722\") " pod="metallb-system/controller-68d546b9d8-74qhz" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.622793 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55mfq\" (UniqueName: \"kubernetes.io/projected/8600909e-f2a1-4472-89ca-275d753fc807-kube-api-access-55mfq\") pod \"speaker-m6ncq\" (UID: \"8600909e-f2a1-4472-89ca-275d753fc807\") " pod="metallb-system/speaker-m6ncq" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.622827 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18786238-f18b-4b0a-b54d-e838c1092722-metrics-certs\") pod \"controller-68d546b9d8-74qhz\" (UID: \"18786238-f18b-4b0a-b54d-e838c1092722\") " pod="metallb-system/controller-68d546b9d8-74qhz" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.622852 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8cd364fe-896d-461f-b118-f757d17c3d1b-cert\") pod \"frr-k8s-webhook-server-64bf5d555-pljrz\" (UID: \"8cd364fe-896d-461f-b118-f757d17c3d1b\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pljrz" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.622876 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8600909e-f2a1-4472-89ca-275d753fc807-memberlist\") pod \"speaker-m6ncq\" (UID: \"8600909e-f2a1-4472-89ca-275d753fc807\") " pod="metallb-system/speaker-m6ncq" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.622904 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdhxp\" (UniqueName: \"kubernetes.io/projected/8cd364fe-896d-461f-b118-f757d17c3d1b-kube-api-access-gdhxp\") pod \"frr-k8s-webhook-server-64bf5d555-pljrz\" (UID: \"8cd364fe-896d-461f-b118-f757d17c3d1b\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pljrz" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.622985 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/853c1c50-69e1-47ca-9951-c75e52e11cc2-frr-startup\") pod \"frr-k8s-jlmg7\" (UID: \"853c1c50-69e1-47ca-9951-c75e52e11cc2\") " pod="metallb-system/frr-k8s-jlmg7" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.623033 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/853c1c50-69e1-47ca-9951-c75e52e11cc2-reloader\") pod \"frr-k8s-jlmg7\" (UID: \"853c1c50-69e1-47ca-9951-c75e52e11cc2\") " pod="metallb-system/frr-k8s-jlmg7" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.623077 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hhck\" (UniqueName: \"kubernetes.io/projected/853c1c50-69e1-47ca-9951-c75e52e11cc2-kube-api-access-2hhck\") pod \"frr-k8s-jlmg7\" (UID: \"853c1c50-69e1-47ca-9951-c75e52e11cc2\") " pod="metallb-system/frr-k8s-jlmg7" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.723674 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8600909e-f2a1-4472-89ca-275d753fc807-memberlist\") pod \"speaker-m6ncq\" (UID: \"8600909e-f2a1-4472-89ca-275d753fc807\") " pod="metallb-system/speaker-m6ncq" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.723721 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdhxp\" (UniqueName: \"kubernetes.io/projected/8cd364fe-896d-461f-b118-f757d17c3d1b-kube-api-access-gdhxp\") pod \"frr-k8s-webhook-server-64bf5d555-pljrz\" (UID: \"8cd364fe-896d-461f-b118-f757d17c3d1b\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pljrz" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.723743 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/853c1c50-69e1-47ca-9951-c75e52e11cc2-frr-startup\") pod \"frr-k8s-jlmg7\" (UID: \"853c1c50-69e1-47ca-9951-c75e52e11cc2\") " pod="metallb-system/frr-k8s-jlmg7" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.723763 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/853c1c50-69e1-47ca-9951-c75e52e11cc2-reloader\") pod \"frr-k8s-jlmg7\" (UID: \"853c1c50-69e1-47ca-9951-c75e52e11cc2\") " pod="metallb-system/frr-k8s-jlmg7" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.723782 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hhck\" (UniqueName: \"kubernetes.io/projected/853c1c50-69e1-47ca-9951-c75e52e11cc2-kube-api-access-2hhck\") pod \"frr-k8s-jlmg7\" (UID: \"853c1c50-69e1-47ca-9951-c75e52e11cc2\") " pod="metallb-system/frr-k8s-jlmg7" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.723800 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/853c1c50-69e1-47ca-9951-c75e52e11cc2-frr-sockets\") pod \"frr-k8s-jlmg7\" (UID: \"853c1c50-69e1-47ca-9951-c75e52e11cc2\") " pod="metallb-system/frr-k8s-jlmg7" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.723816 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/853c1c50-69e1-47ca-9951-c75e52e11cc2-frr-conf\") pod \"frr-k8s-jlmg7\" (UID: \"853c1c50-69e1-47ca-9951-c75e52e11cc2\") " pod="metallb-system/frr-k8s-jlmg7" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.723831 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/8600909e-f2a1-4472-89ca-275d753fc807-metallb-excludel2\") pod \"speaker-m6ncq\" (UID: \"8600909e-f2a1-4472-89ca-275d753fc807\") " pod="metallb-system/speaker-m6ncq" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.723856 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/853c1c50-69e1-47ca-9951-c75e52e11cc2-metrics-certs\") pod \"frr-k8s-jlmg7\" (UID: \"853c1c50-69e1-47ca-9951-c75e52e11cc2\") " pod="metallb-system/frr-k8s-jlmg7" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.723870 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8600909e-f2a1-4472-89ca-275d753fc807-metrics-certs\") pod \"speaker-m6ncq\" (UID: \"8600909e-f2a1-4472-89ca-275d753fc807\") " pod="metallb-system/speaker-m6ncq" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.723886 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/853c1c50-69e1-47ca-9951-c75e52e11cc2-metrics\") pod \"frr-k8s-jlmg7\" (UID: \"853c1c50-69e1-47ca-9951-c75e52e11cc2\") " pod="metallb-system/frr-k8s-jlmg7" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.723900 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7brt\" (UniqueName: \"kubernetes.io/projected/18786238-f18b-4b0a-b54d-e838c1092722-kube-api-access-d7brt\") pod \"controller-68d546b9d8-74qhz\" (UID: \"18786238-f18b-4b0a-b54d-e838c1092722\") " pod="metallb-system/controller-68d546b9d8-74qhz" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.723919 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/18786238-f18b-4b0a-b54d-e838c1092722-cert\") pod \"controller-68d546b9d8-74qhz\" (UID: \"18786238-f18b-4b0a-b54d-e838c1092722\") " pod="metallb-system/controller-68d546b9d8-74qhz" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.723933 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55mfq\" (UniqueName: \"kubernetes.io/projected/8600909e-f2a1-4472-89ca-275d753fc807-kube-api-access-55mfq\") pod \"speaker-m6ncq\" (UID: \"8600909e-f2a1-4472-89ca-275d753fc807\") " pod="metallb-system/speaker-m6ncq" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.723950 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18786238-f18b-4b0a-b54d-e838c1092722-metrics-certs\") pod \"controller-68d546b9d8-74qhz\" (UID: \"18786238-f18b-4b0a-b54d-e838c1092722\") " pod="metallb-system/controller-68d546b9d8-74qhz" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.723971 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8cd364fe-896d-461f-b118-f757d17c3d1b-cert\") pod \"frr-k8s-webhook-server-64bf5d555-pljrz\" (UID: \"8cd364fe-896d-461f-b118-f757d17c3d1b\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pljrz" Oct 02 10:02:05 crc kubenswrapper[4934]: E1002 10:02:05.724069 4934 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Oct 02 10:02:05 crc kubenswrapper[4934]: E1002 10:02:05.724119 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8cd364fe-896d-461f-b118-f757d17c3d1b-cert podName:8cd364fe-896d-461f-b118-f757d17c3d1b nodeName:}" failed. No retries permitted until 2025-10-02 10:02:06.224103187 +0000 UTC m=+797.976744699 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8cd364fe-896d-461f-b118-f757d17c3d1b-cert") pod "frr-k8s-webhook-server-64bf5d555-pljrz" (UID: "8cd364fe-896d-461f-b118-f757d17c3d1b") : secret "frr-k8s-webhook-server-cert" not found Oct 02 10:02:05 crc kubenswrapper[4934]: E1002 10:02:05.724354 4934 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 02 10:02:05 crc kubenswrapper[4934]: E1002 10:02:05.724378 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8600909e-f2a1-4472-89ca-275d753fc807-memberlist podName:8600909e-f2a1-4472-89ca-275d753fc807 nodeName:}" failed. No retries permitted until 2025-10-02 10:02:06.224370654 +0000 UTC m=+797.977012176 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/8600909e-f2a1-4472-89ca-275d753fc807-memberlist") pod "speaker-m6ncq" (UID: "8600909e-f2a1-4472-89ca-275d753fc807") : secret "metallb-memberlist" not found Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.725387 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/853c1c50-69e1-47ca-9951-c75e52e11cc2-frr-startup\") pod \"frr-k8s-jlmg7\" (UID: \"853c1c50-69e1-47ca-9951-c75e52e11cc2\") " pod="metallb-system/frr-k8s-jlmg7" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.725657 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/853c1c50-69e1-47ca-9951-c75e52e11cc2-reloader\") pod \"frr-k8s-jlmg7\" (UID: \"853c1c50-69e1-47ca-9951-c75e52e11cc2\") " pod="metallb-system/frr-k8s-jlmg7" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.725982 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/853c1c50-69e1-47ca-9951-c75e52e11cc2-frr-sockets\") pod \"frr-k8s-jlmg7\" (UID: \"853c1c50-69e1-47ca-9951-c75e52e11cc2\") " pod="metallb-system/frr-k8s-jlmg7" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.726163 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/853c1c50-69e1-47ca-9951-c75e52e11cc2-frr-conf\") pod \"frr-k8s-jlmg7\" (UID: \"853c1c50-69e1-47ca-9951-c75e52e11cc2\") " pod="metallb-system/frr-k8s-jlmg7" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.726727 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/8600909e-f2a1-4472-89ca-275d753fc807-metallb-excludel2\") pod \"speaker-m6ncq\" (UID: \"8600909e-f2a1-4472-89ca-275d753fc807\") " pod="metallb-system/speaker-m6ncq" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.728123 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/853c1c50-69e1-47ca-9951-c75e52e11cc2-metrics\") pod \"frr-k8s-jlmg7\" (UID: \"853c1c50-69e1-47ca-9951-c75e52e11cc2\") " pod="metallb-system/frr-k8s-jlmg7" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.731897 4934 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.732322 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18786238-f18b-4b0a-b54d-e838c1092722-metrics-certs\") pod \"controller-68d546b9d8-74qhz\" (UID: \"18786238-f18b-4b0a-b54d-e838c1092722\") " pod="metallb-system/controller-68d546b9d8-74qhz" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.733987 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8600909e-f2a1-4472-89ca-275d753fc807-metrics-certs\") pod \"speaker-m6ncq\" (UID: \"8600909e-f2a1-4472-89ca-275d753fc807\") " pod="metallb-system/speaker-m6ncq" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.735032 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/853c1c50-69e1-47ca-9951-c75e52e11cc2-metrics-certs\") pod \"frr-k8s-jlmg7\" (UID: \"853c1c50-69e1-47ca-9951-c75e52e11cc2\") " pod="metallb-system/frr-k8s-jlmg7" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.741419 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdhxp\" (UniqueName: \"kubernetes.io/projected/8cd364fe-896d-461f-b118-f757d17c3d1b-kube-api-access-gdhxp\") pod \"frr-k8s-webhook-server-64bf5d555-pljrz\" (UID: \"8cd364fe-896d-461f-b118-f757d17c3d1b\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pljrz" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.741800 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/18786238-f18b-4b0a-b54d-e838c1092722-cert\") pod \"controller-68d546b9d8-74qhz\" (UID: \"18786238-f18b-4b0a-b54d-e838c1092722\") " pod="metallb-system/controller-68d546b9d8-74qhz" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.750138 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55mfq\" (UniqueName: \"kubernetes.io/projected/8600909e-f2a1-4472-89ca-275d753fc807-kube-api-access-55mfq\") pod \"speaker-m6ncq\" (UID: \"8600909e-f2a1-4472-89ca-275d753fc807\") " pod="metallb-system/speaker-m6ncq" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.750388 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7brt\" (UniqueName: \"kubernetes.io/projected/18786238-f18b-4b0a-b54d-e838c1092722-kube-api-access-d7brt\") pod \"controller-68d546b9d8-74qhz\" (UID: \"18786238-f18b-4b0a-b54d-e838c1092722\") " pod="metallb-system/controller-68d546b9d8-74qhz" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.756548 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hhck\" (UniqueName: \"kubernetes.io/projected/853c1c50-69e1-47ca-9951-c75e52e11cc2-kube-api-access-2hhck\") pod \"frr-k8s-jlmg7\" (UID: \"853c1c50-69e1-47ca-9951-c75e52e11cc2\") " pod="metallb-system/frr-k8s-jlmg7" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.772229 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-jlmg7" Oct 02 10:02:05 crc kubenswrapper[4934]: I1002 10:02:05.895444 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-74qhz" Oct 02 10:02:06 crc kubenswrapper[4934]: I1002 10:02:06.197693 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jlmg7" event={"ID":"853c1c50-69e1-47ca-9951-c75e52e11cc2","Type":"ContainerStarted","Data":"146e95a22245c40d34b3adb3ce89f40183883e67e969658e8c7103c063d6abb3"} Oct 02 10:02:06 crc kubenswrapper[4934]: I1002 10:02:06.200148 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6hkdz" Oct 02 10:02:06 crc kubenswrapper[4934]: I1002 10:02:06.200212 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6hkdz" Oct 02 10:02:06 crc kubenswrapper[4934]: I1002 10:02:06.230869 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8cd364fe-896d-461f-b118-f757d17c3d1b-cert\") pod \"frr-k8s-webhook-server-64bf5d555-pljrz\" (UID: \"8cd364fe-896d-461f-b118-f757d17c3d1b\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pljrz" Oct 02 10:02:06 crc kubenswrapper[4934]: I1002 10:02:06.230945 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8600909e-f2a1-4472-89ca-275d753fc807-memberlist\") pod \"speaker-m6ncq\" (UID: \"8600909e-f2a1-4472-89ca-275d753fc807\") " pod="metallb-system/speaker-m6ncq" Oct 02 10:02:06 crc kubenswrapper[4934]: E1002 10:02:06.231096 4934 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 02 10:02:06 crc kubenswrapper[4934]: E1002 10:02:06.231160 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8600909e-f2a1-4472-89ca-275d753fc807-memberlist podName:8600909e-f2a1-4472-89ca-275d753fc807 nodeName:}" failed. No retries permitted until 2025-10-02 10:02:07.231141458 +0000 UTC m=+798.983782980 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/8600909e-f2a1-4472-89ca-275d753fc807-memberlist") pod "speaker-m6ncq" (UID: "8600909e-f2a1-4472-89ca-275d753fc807") : secret "metallb-memberlist" not found Oct 02 10:02:06 crc kubenswrapper[4934]: I1002 10:02:06.237405 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8cd364fe-896d-461f-b118-f757d17c3d1b-cert\") pod \"frr-k8s-webhook-server-64bf5d555-pljrz\" (UID: \"8cd364fe-896d-461f-b118-f757d17c3d1b\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pljrz" Oct 02 10:02:06 crc kubenswrapper[4934]: I1002 10:02:06.243705 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6hkdz" Oct 02 10:02:06 crc kubenswrapper[4934]: I1002 10:02:06.279693 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-74qhz"] Oct 02 10:02:06 crc kubenswrapper[4934]: W1002 10:02:06.281633 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18786238_f18b_4b0a_b54d_e838c1092722.slice/crio-077f25f850f5e493ed93aa4b36d6d4822b5ff29c979191d381c93c76e9e3f9e7 WatchSource:0}: Error finding container 077f25f850f5e493ed93aa4b36d6d4822b5ff29c979191d381c93c76e9e3f9e7: Status 404 returned error can't find the container with id 077f25f850f5e493ed93aa4b36d6d4822b5ff29c979191d381c93c76e9e3f9e7 Oct 02 10:02:06 crc kubenswrapper[4934]: I1002 10:02:06.362334 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pljrz" Oct 02 10:02:06 crc kubenswrapper[4934]: I1002 10:02:06.818432 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-pljrz"] Oct 02 10:02:06 crc kubenswrapper[4934]: W1002 10:02:06.825351 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cd364fe_896d_461f_b118_f757d17c3d1b.slice/crio-7f6955f241ef379601f39190de5d1ea736e6d770c6c66520a7395f9e9b07f206 WatchSource:0}: Error finding container 7f6955f241ef379601f39190de5d1ea736e6d770c6c66520a7395f9e9b07f206: Status 404 returned error can't find the container with id 7f6955f241ef379601f39190de5d1ea736e6d770c6c66520a7395f9e9b07f206 Oct 02 10:02:07 crc kubenswrapper[4934]: I1002 10:02:07.205211 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pljrz" event={"ID":"8cd364fe-896d-461f-b118-f757d17c3d1b","Type":"ContainerStarted","Data":"7f6955f241ef379601f39190de5d1ea736e6d770c6c66520a7395f9e9b07f206"} Oct 02 10:02:07 crc kubenswrapper[4934]: I1002 10:02:07.207729 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-74qhz" event={"ID":"18786238-f18b-4b0a-b54d-e838c1092722","Type":"ContainerStarted","Data":"7be2a82046bdb7d9009a03fb2b553efdf5ed542770d8f041e7f05e8487cdb7ce"} Oct 02 10:02:07 crc kubenswrapper[4934]: I1002 10:02:07.207787 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-74qhz" event={"ID":"18786238-f18b-4b0a-b54d-e838c1092722","Type":"ContainerStarted","Data":"b39ea7d98b44fb0365ff0a6bad251cc3bd3afe04cd8bb95b88a1d11ad5d67f85"} Oct 02 10:02:07 crc kubenswrapper[4934]: I1002 10:02:07.207803 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-74qhz" event={"ID":"18786238-f18b-4b0a-b54d-e838c1092722","Type":"ContainerStarted","Data":"077f25f850f5e493ed93aa4b36d6d4822b5ff29c979191d381c93c76e9e3f9e7"} Oct 02 10:02:07 crc kubenswrapper[4934]: I1002 10:02:07.225654 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-74qhz" podStartSLOduration=2.22563651 podStartE2EDuration="2.22563651s" podCreationTimestamp="2025-10-02 10:02:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:02:07.22274909 +0000 UTC m=+798.975390612" watchObservedRunningTime="2025-10-02 10:02:07.22563651 +0000 UTC m=+798.978278032" Oct 02 10:02:07 crc kubenswrapper[4934]: I1002 10:02:07.242226 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8600909e-f2a1-4472-89ca-275d753fc807-memberlist\") pod \"speaker-m6ncq\" (UID: \"8600909e-f2a1-4472-89ca-275d753fc807\") " pod="metallb-system/speaker-m6ncq" Oct 02 10:02:07 crc kubenswrapper[4934]: I1002 10:02:07.249665 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8600909e-f2a1-4472-89ca-275d753fc807-memberlist\") pod \"speaker-m6ncq\" (UID: \"8600909e-f2a1-4472-89ca-275d753fc807\") " pod="metallb-system/speaker-m6ncq" Oct 02 10:02:07 crc kubenswrapper[4934]: I1002 10:02:07.274033 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6hkdz" Oct 02 10:02:07 crc kubenswrapper[4934]: I1002 10:02:07.376397 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-m6ncq" Oct 02 10:02:07 crc kubenswrapper[4934]: W1002 10:02:07.412030 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8600909e_f2a1_4472_89ca_275d753fc807.slice/crio-fc62c650c33d2507015c2b2af618ab9041fa400748e5e4bed7d890b38fa8dd55 WatchSource:0}: Error finding container fc62c650c33d2507015c2b2af618ab9041fa400748e5e4bed7d890b38fa8dd55: Status 404 returned error can't find the container with id fc62c650c33d2507015c2b2af618ab9041fa400748e5e4bed7d890b38fa8dd55 Oct 02 10:02:08 crc kubenswrapper[4934]: I1002 10:02:08.228124 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-m6ncq" event={"ID":"8600909e-f2a1-4472-89ca-275d753fc807","Type":"ContainerStarted","Data":"b12460392ae5563316726065e611e5c52b3b1953a9fc4c4dabca1a250c6b51c5"} Oct 02 10:02:08 crc kubenswrapper[4934]: I1002 10:02:08.228386 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-74qhz" Oct 02 10:02:08 crc kubenswrapper[4934]: I1002 10:02:08.228399 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-m6ncq" event={"ID":"8600909e-f2a1-4472-89ca-275d753fc807","Type":"ContainerStarted","Data":"77ed825630012702350d5b88698f4e4662fb03a35983b07cd6d596cb70db39e0"} Oct 02 10:02:08 crc kubenswrapper[4934]: I1002 10:02:08.228408 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-m6ncq" event={"ID":"8600909e-f2a1-4472-89ca-275d753fc807","Type":"ContainerStarted","Data":"fc62c650c33d2507015c2b2af618ab9041fa400748e5e4bed7d890b38fa8dd55"} Oct 02 10:02:08 crc kubenswrapper[4934]: I1002 10:02:08.228726 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-m6ncq" Oct 02 10:02:08 crc kubenswrapper[4934]: I1002 10:02:08.246513 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-m6ncq" podStartSLOduration=3.24649333 podStartE2EDuration="3.24649333s" podCreationTimestamp="2025-10-02 10:02:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:02:08.243723153 +0000 UTC m=+799.996364685" watchObservedRunningTime="2025-10-02 10:02:08.24649333 +0000 UTC m=+799.999134852" Oct 02 10:02:08 crc kubenswrapper[4934]: I1002 10:02:08.661810 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6hkdz"] Oct 02 10:02:09 crc kubenswrapper[4934]: I1002 10:02:09.233390 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6hkdz" podUID="5be6c267-644a-4535-b143-0f70ed2e7efd" containerName="registry-server" containerID="cri-o://5cee34364539c42a7dbce13d979203513de088356f120f9cd21e245eba1be240" gracePeriod=2 Oct 02 10:02:09 crc kubenswrapper[4934]: I1002 10:02:09.659980 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6hkdz" Oct 02 10:02:09 crc kubenswrapper[4934]: I1002 10:02:09.680613 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sg5g9\" (UniqueName: \"kubernetes.io/projected/5be6c267-644a-4535-b143-0f70ed2e7efd-kube-api-access-sg5g9\") pod \"5be6c267-644a-4535-b143-0f70ed2e7efd\" (UID: \"5be6c267-644a-4535-b143-0f70ed2e7efd\") " Oct 02 10:02:09 crc kubenswrapper[4934]: I1002 10:02:09.680699 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5be6c267-644a-4535-b143-0f70ed2e7efd-catalog-content\") pod \"5be6c267-644a-4535-b143-0f70ed2e7efd\" (UID: \"5be6c267-644a-4535-b143-0f70ed2e7efd\") " Oct 02 10:02:09 crc kubenswrapper[4934]: I1002 10:02:09.680753 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5be6c267-644a-4535-b143-0f70ed2e7efd-utilities\") pod \"5be6c267-644a-4535-b143-0f70ed2e7efd\" (UID: \"5be6c267-644a-4535-b143-0f70ed2e7efd\") " Oct 02 10:02:09 crc kubenswrapper[4934]: I1002 10:02:09.684087 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5be6c267-644a-4535-b143-0f70ed2e7efd-utilities" (OuterVolumeSpecName: "utilities") pod "5be6c267-644a-4535-b143-0f70ed2e7efd" (UID: "5be6c267-644a-4535-b143-0f70ed2e7efd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:02:09 crc kubenswrapper[4934]: I1002 10:02:09.697435 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5be6c267-644a-4535-b143-0f70ed2e7efd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5be6c267-644a-4535-b143-0f70ed2e7efd" (UID: "5be6c267-644a-4535-b143-0f70ed2e7efd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:02:09 crc kubenswrapper[4934]: I1002 10:02:09.702997 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5be6c267-644a-4535-b143-0f70ed2e7efd-kube-api-access-sg5g9" (OuterVolumeSpecName: "kube-api-access-sg5g9") pod "5be6c267-644a-4535-b143-0f70ed2e7efd" (UID: "5be6c267-644a-4535-b143-0f70ed2e7efd"). InnerVolumeSpecName "kube-api-access-sg5g9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:02:09 crc kubenswrapper[4934]: I1002 10:02:09.781680 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sg5g9\" (UniqueName: \"kubernetes.io/projected/5be6c267-644a-4535-b143-0f70ed2e7efd-kube-api-access-sg5g9\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:09 crc kubenswrapper[4934]: I1002 10:02:09.781714 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5be6c267-644a-4535-b143-0f70ed2e7efd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:09 crc kubenswrapper[4934]: I1002 10:02:09.781723 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5be6c267-644a-4535-b143-0f70ed2e7efd-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:10 crc kubenswrapper[4934]: I1002 10:02:10.240523 4934 generic.go:334] "Generic (PLEG): container finished" podID="5be6c267-644a-4535-b143-0f70ed2e7efd" containerID="5cee34364539c42a7dbce13d979203513de088356f120f9cd21e245eba1be240" exitCode=0 Oct 02 10:02:10 crc kubenswrapper[4934]: I1002 10:02:10.240571 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6hkdz" Oct 02 10:02:10 crc kubenswrapper[4934]: I1002 10:02:10.240595 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hkdz" event={"ID":"5be6c267-644a-4535-b143-0f70ed2e7efd","Type":"ContainerDied","Data":"5cee34364539c42a7dbce13d979203513de088356f120f9cd21e245eba1be240"} Oct 02 10:02:10 crc kubenswrapper[4934]: I1002 10:02:10.240999 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hkdz" event={"ID":"5be6c267-644a-4535-b143-0f70ed2e7efd","Type":"ContainerDied","Data":"084a43750c18c053d28e7ef7d580d1425ae2344beb885ebbcaaa142296516ffc"} Oct 02 10:02:10 crc kubenswrapper[4934]: I1002 10:02:10.241018 4934 scope.go:117] "RemoveContainer" containerID="5cee34364539c42a7dbce13d979203513de088356f120f9cd21e245eba1be240" Oct 02 10:02:10 crc kubenswrapper[4934]: I1002 10:02:10.266072 4934 scope.go:117] "RemoveContainer" containerID="ed11ac14839ae582fa8a07660889121a912b2747f11f2323c636c47866b576db" Oct 02 10:02:10 crc kubenswrapper[4934]: I1002 10:02:10.271910 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6hkdz"] Oct 02 10:02:10 crc kubenswrapper[4934]: I1002 10:02:10.275272 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6hkdz"] Oct 02 10:02:10 crc kubenswrapper[4934]: I1002 10:02:10.306187 4934 scope.go:117] "RemoveContainer" containerID="605e2b7746b83422eaaa5eb57ebe09949a45d4da05f770f12d9bf09b01c3324f" Oct 02 10:02:10 crc kubenswrapper[4934]: I1002 10:02:10.328695 4934 scope.go:117] "RemoveContainer" containerID="5cee34364539c42a7dbce13d979203513de088356f120f9cd21e245eba1be240" Oct 02 10:02:10 crc kubenswrapper[4934]: E1002 10:02:10.329184 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cee34364539c42a7dbce13d979203513de088356f120f9cd21e245eba1be240\": container with ID starting with 5cee34364539c42a7dbce13d979203513de088356f120f9cd21e245eba1be240 not found: ID does not exist" containerID="5cee34364539c42a7dbce13d979203513de088356f120f9cd21e245eba1be240" Oct 02 10:02:10 crc kubenswrapper[4934]: I1002 10:02:10.329222 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cee34364539c42a7dbce13d979203513de088356f120f9cd21e245eba1be240"} err="failed to get container status \"5cee34364539c42a7dbce13d979203513de088356f120f9cd21e245eba1be240\": rpc error: code = NotFound desc = could not find container \"5cee34364539c42a7dbce13d979203513de088356f120f9cd21e245eba1be240\": container with ID starting with 5cee34364539c42a7dbce13d979203513de088356f120f9cd21e245eba1be240 not found: ID does not exist" Oct 02 10:02:10 crc kubenswrapper[4934]: I1002 10:02:10.329246 4934 scope.go:117] "RemoveContainer" containerID="ed11ac14839ae582fa8a07660889121a912b2747f11f2323c636c47866b576db" Oct 02 10:02:10 crc kubenswrapper[4934]: E1002 10:02:10.329842 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed11ac14839ae582fa8a07660889121a912b2747f11f2323c636c47866b576db\": container with ID starting with ed11ac14839ae582fa8a07660889121a912b2747f11f2323c636c47866b576db not found: ID does not exist" containerID="ed11ac14839ae582fa8a07660889121a912b2747f11f2323c636c47866b576db" Oct 02 10:02:10 crc kubenswrapper[4934]: I1002 10:02:10.329885 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed11ac14839ae582fa8a07660889121a912b2747f11f2323c636c47866b576db"} err="failed to get container status \"ed11ac14839ae582fa8a07660889121a912b2747f11f2323c636c47866b576db\": rpc error: code = NotFound desc = could not find container \"ed11ac14839ae582fa8a07660889121a912b2747f11f2323c636c47866b576db\": container with ID starting with ed11ac14839ae582fa8a07660889121a912b2747f11f2323c636c47866b576db not found: ID does not exist" Oct 02 10:02:10 crc kubenswrapper[4934]: I1002 10:02:10.329915 4934 scope.go:117] "RemoveContainer" containerID="605e2b7746b83422eaaa5eb57ebe09949a45d4da05f770f12d9bf09b01c3324f" Oct 02 10:02:10 crc kubenswrapper[4934]: E1002 10:02:10.330698 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"605e2b7746b83422eaaa5eb57ebe09949a45d4da05f770f12d9bf09b01c3324f\": container with ID starting with 605e2b7746b83422eaaa5eb57ebe09949a45d4da05f770f12d9bf09b01c3324f not found: ID does not exist" containerID="605e2b7746b83422eaaa5eb57ebe09949a45d4da05f770f12d9bf09b01c3324f" Oct 02 10:02:10 crc kubenswrapper[4934]: I1002 10:02:10.330725 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"605e2b7746b83422eaaa5eb57ebe09949a45d4da05f770f12d9bf09b01c3324f"} err="failed to get container status \"605e2b7746b83422eaaa5eb57ebe09949a45d4da05f770f12d9bf09b01c3324f\": rpc error: code = NotFound desc = could not find container \"605e2b7746b83422eaaa5eb57ebe09949a45d4da05f770f12d9bf09b01c3324f\": container with ID starting with 605e2b7746b83422eaaa5eb57ebe09949a45d4da05f770f12d9bf09b01c3324f not found: ID does not exist" Oct 02 10:02:10 crc kubenswrapper[4934]: I1002 10:02:10.921897 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5be6c267-644a-4535-b143-0f70ed2e7efd" path="/var/lib/kubelet/pods/5be6c267-644a-4535-b143-0f70ed2e7efd/volumes" Oct 02 10:02:14 crc kubenswrapper[4934]: I1002 10:02:14.274405 4934 generic.go:334] "Generic (PLEG): container finished" podID="853c1c50-69e1-47ca-9951-c75e52e11cc2" containerID="26084c7dfcab49429a35f1c5265c1c16aaf276e6e05f1a471eca6910ad1a4d64" exitCode=0 Oct 02 10:02:14 crc kubenswrapper[4934]: I1002 10:02:14.274466 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jlmg7" event={"ID":"853c1c50-69e1-47ca-9951-c75e52e11cc2","Type":"ContainerDied","Data":"26084c7dfcab49429a35f1c5265c1c16aaf276e6e05f1a471eca6910ad1a4d64"} Oct 02 10:02:14 crc kubenswrapper[4934]: I1002 10:02:14.276832 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pljrz" event={"ID":"8cd364fe-896d-461f-b118-f757d17c3d1b","Type":"ContainerStarted","Data":"5c53fb74aea6a32cb5359893c20408a0bbbaf1dc7ea7c26d64d02135582e96e1"} Oct 02 10:02:14 crc kubenswrapper[4934]: I1002 10:02:14.277033 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pljrz" Oct 02 10:02:15 crc kubenswrapper[4934]: I1002 10:02:15.285103 4934 generic.go:334] "Generic (PLEG): container finished" podID="853c1c50-69e1-47ca-9951-c75e52e11cc2" containerID="0130fe2211ea1fa988ab143876b5712dbd4001890587689a0c4b8439b2fc80de" exitCode=0 Oct 02 10:02:15 crc kubenswrapper[4934]: I1002 10:02:15.285188 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jlmg7" event={"ID":"853c1c50-69e1-47ca-9951-c75e52e11cc2","Type":"ContainerDied","Data":"0130fe2211ea1fa988ab143876b5712dbd4001890587689a0c4b8439b2fc80de"} Oct 02 10:02:15 crc kubenswrapper[4934]: I1002 10:02:15.313152 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pljrz" podStartSLOduration=3.9613919920000003 podStartE2EDuration="10.31313551s" podCreationTimestamp="2025-10-02 10:02:05 +0000 UTC" firstStartedPulling="2025-10-02 10:02:06.827912267 +0000 UTC m=+798.580553789" lastFinishedPulling="2025-10-02 10:02:13.179655765 +0000 UTC m=+804.932297307" observedRunningTime="2025-10-02 10:02:14.349804928 +0000 UTC m=+806.102446480" watchObservedRunningTime="2025-10-02 10:02:15.31313551 +0000 UTC m=+807.065777032" Oct 02 10:02:16 crc kubenswrapper[4934]: I1002 10:02:16.296342 4934 generic.go:334] "Generic (PLEG): container finished" podID="853c1c50-69e1-47ca-9951-c75e52e11cc2" containerID="9535234ea60e31e9b837e1688de171723b82494829e1bcb69cea7513869b5ce5" exitCode=0 Oct 02 10:02:16 crc kubenswrapper[4934]: I1002 10:02:16.296397 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jlmg7" event={"ID":"853c1c50-69e1-47ca-9951-c75e52e11cc2","Type":"ContainerDied","Data":"9535234ea60e31e9b837e1688de171723b82494829e1bcb69cea7513869b5ce5"} Oct 02 10:02:17 crc kubenswrapper[4934]: I1002 10:02:17.308436 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jlmg7" event={"ID":"853c1c50-69e1-47ca-9951-c75e52e11cc2","Type":"ContainerStarted","Data":"78ee72706e134ad07c0623c677be46846b42263b16f46d585c62bce0b49f4d70"} Oct 02 10:02:17 crc kubenswrapper[4934]: I1002 10:02:17.308793 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jlmg7" event={"ID":"853c1c50-69e1-47ca-9951-c75e52e11cc2","Type":"ContainerStarted","Data":"57949941b3875960a3ab26567a854ba2d3f9771ac54646d1b3c56f21900e94e0"} Oct 02 10:02:17 crc kubenswrapper[4934]: I1002 10:02:17.308804 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jlmg7" event={"ID":"853c1c50-69e1-47ca-9951-c75e52e11cc2","Type":"ContainerStarted","Data":"0cddf928b29303fcb3707e8d0ac24764550646e2e956d9b2a65c762c0dfa9907"} Oct 02 10:02:17 crc kubenswrapper[4934]: I1002 10:02:17.308814 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jlmg7" event={"ID":"853c1c50-69e1-47ca-9951-c75e52e11cc2","Type":"ContainerStarted","Data":"f4ea0bb0882b24ae75fa52e2a586c4e61c102f932690f633b648e7912c7f6f14"} Oct 02 10:02:17 crc kubenswrapper[4934]: I1002 10:02:17.308823 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jlmg7" event={"ID":"853c1c50-69e1-47ca-9951-c75e52e11cc2","Type":"ContainerStarted","Data":"bb6fd491e8249e35aea9abaa4b6af935038e94c05e2fb6f3ceb106053066ed6f"} Oct 02 10:02:17 crc kubenswrapper[4934]: I1002 10:02:17.308839 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-jlmg7" Oct 02 10:02:17 crc kubenswrapper[4934]: I1002 10:02:17.308848 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jlmg7" event={"ID":"853c1c50-69e1-47ca-9951-c75e52e11cc2","Type":"ContainerStarted","Data":"ebe8b724aff8c1d2f964f6182441191022b144006e802c1699454ca6ecd7530f"} Oct 02 10:02:17 crc kubenswrapper[4934]: I1002 10:02:17.335524 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-jlmg7" podStartSLOduration=5.055464454 podStartE2EDuration="12.335504056s" podCreationTimestamp="2025-10-02 10:02:05 +0000 UTC" firstStartedPulling="2025-10-02 10:02:05.883066926 +0000 UTC m=+797.635708448" lastFinishedPulling="2025-10-02 10:02:13.163106508 +0000 UTC m=+804.915748050" observedRunningTime="2025-10-02 10:02:17.333823389 +0000 UTC m=+809.086464911" watchObservedRunningTime="2025-10-02 10:02:17.335504056 +0000 UTC m=+809.088145578" Oct 02 10:02:17 crc kubenswrapper[4934]: I1002 10:02:17.384023 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-m6ncq" Oct 02 10:02:19 crc kubenswrapper[4934]: I1002 10:02:19.130138 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw"] Oct 02 10:02:19 crc kubenswrapper[4934]: E1002 10:02:19.130771 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5be6c267-644a-4535-b143-0f70ed2e7efd" containerName="registry-server" Oct 02 10:02:19 crc kubenswrapper[4934]: I1002 10:02:19.130789 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="5be6c267-644a-4535-b143-0f70ed2e7efd" containerName="registry-server" Oct 02 10:02:19 crc kubenswrapper[4934]: E1002 10:02:19.130818 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5be6c267-644a-4535-b143-0f70ed2e7efd" containerName="extract-utilities" Oct 02 10:02:19 crc kubenswrapper[4934]: I1002 10:02:19.130826 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="5be6c267-644a-4535-b143-0f70ed2e7efd" containerName="extract-utilities" Oct 02 10:02:19 crc kubenswrapper[4934]: E1002 10:02:19.130842 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5be6c267-644a-4535-b143-0f70ed2e7efd" containerName="extract-content" Oct 02 10:02:19 crc kubenswrapper[4934]: I1002 10:02:19.130851 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="5be6c267-644a-4535-b143-0f70ed2e7efd" containerName="extract-content" Oct 02 10:02:19 crc kubenswrapper[4934]: I1002 10:02:19.130984 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="5be6c267-644a-4535-b143-0f70ed2e7efd" containerName="registry-server" Oct 02 10:02:19 crc kubenswrapper[4934]: I1002 10:02:19.132044 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw" Oct 02 10:02:19 crc kubenswrapper[4934]: I1002 10:02:19.134980 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 02 10:02:19 crc kubenswrapper[4934]: I1002 10:02:19.168847 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw"] Oct 02 10:02:19 crc kubenswrapper[4934]: I1002 10:02:19.199892 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/537ea528-26ed-4970-a490-e625f7482003-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw\" (UID: \"537ea528-26ed-4970-a490-e625f7482003\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw" Oct 02 10:02:19 crc kubenswrapper[4934]: I1002 10:02:19.200022 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6br2\" (UniqueName: \"kubernetes.io/projected/537ea528-26ed-4970-a490-e625f7482003-kube-api-access-x6br2\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw\" (UID: \"537ea528-26ed-4970-a490-e625f7482003\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw" Oct 02 10:02:19 crc kubenswrapper[4934]: I1002 10:02:19.200080 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/537ea528-26ed-4970-a490-e625f7482003-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw\" (UID: \"537ea528-26ed-4970-a490-e625f7482003\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw" Oct 02 10:02:19 crc kubenswrapper[4934]: I1002 10:02:19.300551 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/537ea528-26ed-4970-a490-e625f7482003-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw\" (UID: \"537ea528-26ed-4970-a490-e625f7482003\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw" Oct 02 10:02:19 crc kubenswrapper[4934]: I1002 10:02:19.300730 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6br2\" (UniqueName: \"kubernetes.io/projected/537ea528-26ed-4970-a490-e625f7482003-kube-api-access-x6br2\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw\" (UID: \"537ea528-26ed-4970-a490-e625f7482003\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw" Oct 02 10:02:19 crc kubenswrapper[4934]: I1002 10:02:19.300803 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/537ea528-26ed-4970-a490-e625f7482003-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw\" (UID: \"537ea528-26ed-4970-a490-e625f7482003\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw" Oct 02 10:02:19 crc kubenswrapper[4934]: I1002 10:02:19.301063 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/537ea528-26ed-4970-a490-e625f7482003-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw\" (UID: \"537ea528-26ed-4970-a490-e625f7482003\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw" Oct 02 10:02:19 crc kubenswrapper[4934]: I1002 10:02:19.301462 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/537ea528-26ed-4970-a490-e625f7482003-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw\" (UID: \"537ea528-26ed-4970-a490-e625f7482003\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw" Oct 02 10:02:19 crc kubenswrapper[4934]: I1002 10:02:19.328784 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6br2\" (UniqueName: \"kubernetes.io/projected/537ea528-26ed-4970-a490-e625f7482003-kube-api-access-x6br2\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw\" (UID: \"537ea528-26ed-4970-a490-e625f7482003\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw" Oct 02 10:02:19 crc kubenswrapper[4934]: I1002 10:02:19.452006 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw" Oct 02 10:02:19 crc kubenswrapper[4934]: I1002 10:02:19.841994 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw"] Oct 02 10:02:19 crc kubenswrapper[4934]: W1002 10:02:19.844978 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod537ea528_26ed_4970_a490_e625f7482003.slice/crio-c046bacb5cd2cd8c961e785902ddeb15b3417da538b126752be9b7433f00b20b WatchSource:0}: Error finding container c046bacb5cd2cd8c961e785902ddeb15b3417da538b126752be9b7433f00b20b: Status 404 returned error can't find the container with id c046bacb5cd2cd8c961e785902ddeb15b3417da538b126752be9b7433f00b20b Oct 02 10:02:20 crc kubenswrapper[4934]: I1002 10:02:20.331780 4934 generic.go:334] "Generic (PLEG): container finished" podID="537ea528-26ed-4970-a490-e625f7482003" containerID="ad36a043da3a3e4e224e67ea54a5788530795d438cab2d79ad69e296949cab6b" exitCode=0 Oct 02 10:02:20 crc kubenswrapper[4934]: I1002 10:02:20.332123 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw" event={"ID":"537ea528-26ed-4970-a490-e625f7482003","Type":"ContainerDied","Data":"ad36a043da3a3e4e224e67ea54a5788530795d438cab2d79ad69e296949cab6b"} Oct 02 10:02:20 crc kubenswrapper[4934]: I1002 10:02:20.332152 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw" event={"ID":"537ea528-26ed-4970-a490-e625f7482003","Type":"ContainerStarted","Data":"c046bacb5cd2cd8c961e785902ddeb15b3417da538b126752be9b7433f00b20b"} Oct 02 10:02:20 crc kubenswrapper[4934]: I1002 10:02:20.773083 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-jlmg7" Oct 02 10:02:20 crc kubenswrapper[4934]: I1002 10:02:20.827955 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-jlmg7" Oct 02 10:02:23 crc kubenswrapper[4934]: I1002 10:02:23.351409 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw" event={"ID":"537ea528-26ed-4970-a490-e625f7482003","Type":"ContainerStarted","Data":"a9a572785aa6722d6d747e23279c8edd6649567b6a99000a2c430051063c7cbc"} Oct 02 10:02:24 crc kubenswrapper[4934]: I1002 10:02:24.358922 4934 generic.go:334] "Generic (PLEG): container finished" podID="537ea528-26ed-4970-a490-e625f7482003" containerID="a9a572785aa6722d6d747e23279c8edd6649567b6a99000a2c430051063c7cbc" exitCode=0 Oct 02 10:02:24 crc kubenswrapper[4934]: I1002 10:02:24.359273 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw" event={"ID":"537ea528-26ed-4970-a490-e625f7482003","Type":"ContainerDied","Data":"a9a572785aa6722d6d747e23279c8edd6649567b6a99000a2c430051063c7cbc"} Oct 02 10:02:25 crc kubenswrapper[4934]: I1002 10:02:25.371994 4934 generic.go:334] "Generic (PLEG): container finished" podID="537ea528-26ed-4970-a490-e625f7482003" containerID="bdf9f7fa71e012393ba92d57c8affd70709ff5d39309c082f9667fec9207976c" exitCode=0 Oct 02 10:02:25 crc kubenswrapper[4934]: I1002 10:02:25.372065 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw" event={"ID":"537ea528-26ed-4970-a490-e625f7482003","Type":"ContainerDied","Data":"bdf9f7fa71e012393ba92d57c8affd70709ff5d39309c082f9667fec9207976c"} Oct 02 10:02:25 crc kubenswrapper[4934]: I1002 10:02:25.898860 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-74qhz" Oct 02 10:02:26 crc kubenswrapper[4934]: I1002 10:02:26.371559 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-pljrz" Oct 02 10:02:26 crc kubenswrapper[4934]: I1002 10:02:26.695154 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw" Oct 02 10:02:26 crc kubenswrapper[4934]: I1002 10:02:26.804968 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/537ea528-26ed-4970-a490-e625f7482003-util\") pod \"537ea528-26ed-4970-a490-e625f7482003\" (UID: \"537ea528-26ed-4970-a490-e625f7482003\") " Oct 02 10:02:26 crc kubenswrapper[4934]: I1002 10:02:26.805026 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6br2\" (UniqueName: \"kubernetes.io/projected/537ea528-26ed-4970-a490-e625f7482003-kube-api-access-x6br2\") pod \"537ea528-26ed-4970-a490-e625f7482003\" (UID: \"537ea528-26ed-4970-a490-e625f7482003\") " Oct 02 10:02:26 crc kubenswrapper[4934]: I1002 10:02:26.805137 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/537ea528-26ed-4970-a490-e625f7482003-bundle\") pod \"537ea528-26ed-4970-a490-e625f7482003\" (UID: \"537ea528-26ed-4970-a490-e625f7482003\") " Oct 02 10:02:26 crc kubenswrapper[4934]: I1002 10:02:26.806464 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/537ea528-26ed-4970-a490-e625f7482003-bundle" (OuterVolumeSpecName: "bundle") pod "537ea528-26ed-4970-a490-e625f7482003" (UID: "537ea528-26ed-4970-a490-e625f7482003"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:02:26 crc kubenswrapper[4934]: I1002 10:02:26.814990 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/537ea528-26ed-4970-a490-e625f7482003-kube-api-access-x6br2" (OuterVolumeSpecName: "kube-api-access-x6br2") pod "537ea528-26ed-4970-a490-e625f7482003" (UID: "537ea528-26ed-4970-a490-e625f7482003"). InnerVolumeSpecName "kube-api-access-x6br2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:02:26 crc kubenswrapper[4934]: I1002 10:02:26.823500 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/537ea528-26ed-4970-a490-e625f7482003-util" (OuterVolumeSpecName: "util") pod "537ea528-26ed-4970-a490-e625f7482003" (UID: "537ea528-26ed-4970-a490-e625f7482003"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:02:26 crc kubenswrapper[4934]: I1002 10:02:26.907060 4934 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/537ea528-26ed-4970-a490-e625f7482003-util\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:26 crc kubenswrapper[4934]: I1002 10:02:26.907116 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6br2\" (UniqueName: \"kubernetes.io/projected/537ea528-26ed-4970-a490-e625f7482003-kube-api-access-x6br2\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:26 crc kubenswrapper[4934]: I1002 10:02:26.907138 4934 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/537ea528-26ed-4970-a490-e625f7482003-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:27 crc kubenswrapper[4934]: I1002 10:02:27.388056 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw" event={"ID":"537ea528-26ed-4970-a490-e625f7482003","Type":"ContainerDied","Data":"c046bacb5cd2cd8c961e785902ddeb15b3417da538b126752be9b7433f00b20b"} Oct 02 10:02:27 crc kubenswrapper[4934]: I1002 10:02:27.388127 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c046bacb5cd2cd8c961e785902ddeb15b3417da538b126752be9b7433f00b20b" Oct 02 10:02:27 crc kubenswrapper[4934]: I1002 10:02:27.388157 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw" Oct 02 10:02:33 crc kubenswrapper[4934]: I1002 10:02:33.657710 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-fmhwf"] Oct 02 10:02:33 crc kubenswrapper[4934]: E1002 10:02:33.658461 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="537ea528-26ed-4970-a490-e625f7482003" containerName="pull" Oct 02 10:02:33 crc kubenswrapper[4934]: I1002 10:02:33.658476 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="537ea528-26ed-4970-a490-e625f7482003" containerName="pull" Oct 02 10:02:33 crc kubenswrapper[4934]: E1002 10:02:33.658491 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="537ea528-26ed-4970-a490-e625f7482003" containerName="extract" Oct 02 10:02:33 crc kubenswrapper[4934]: I1002 10:02:33.658500 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="537ea528-26ed-4970-a490-e625f7482003" containerName="extract" Oct 02 10:02:33 crc kubenswrapper[4934]: E1002 10:02:33.658512 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="537ea528-26ed-4970-a490-e625f7482003" containerName="util" Oct 02 10:02:33 crc kubenswrapper[4934]: I1002 10:02:33.658520 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="537ea528-26ed-4970-a490-e625f7482003" containerName="util" Oct 02 10:02:33 crc kubenswrapper[4934]: I1002 10:02:33.658651 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="537ea528-26ed-4970-a490-e625f7482003" containerName="extract" Oct 02 10:02:33 crc kubenswrapper[4934]: I1002 10:02:33.659135 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-fmhwf" Oct 02 10:02:33 crc kubenswrapper[4934]: I1002 10:02:33.662204 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Oct 02 10:02:33 crc kubenswrapper[4934]: I1002 10:02:33.662557 4934 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-6d44s" Oct 02 10:02:33 crc kubenswrapper[4934]: I1002 10:02:33.664454 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Oct 02 10:02:33 crc kubenswrapper[4934]: I1002 10:02:33.677303 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-fmhwf"] Oct 02 10:02:33 crc kubenswrapper[4934]: I1002 10:02:33.798246 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnqnv\" (UniqueName: \"kubernetes.io/projected/82ee4886-7435-48d7-bcbb-18b5c2f68758-kube-api-access-vnqnv\") pod \"cert-manager-operator-controller-manager-57cd46d6d-fmhwf\" (UID: \"82ee4886-7435-48d7-bcbb-18b5c2f68758\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-fmhwf" Oct 02 10:02:33 crc kubenswrapper[4934]: I1002 10:02:33.899610 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnqnv\" (UniqueName: \"kubernetes.io/projected/82ee4886-7435-48d7-bcbb-18b5c2f68758-kube-api-access-vnqnv\") pod \"cert-manager-operator-controller-manager-57cd46d6d-fmhwf\" (UID: \"82ee4886-7435-48d7-bcbb-18b5c2f68758\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-fmhwf" Oct 02 10:02:33 crc kubenswrapper[4934]: I1002 10:02:33.925615 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnqnv\" (UniqueName: \"kubernetes.io/projected/82ee4886-7435-48d7-bcbb-18b5c2f68758-kube-api-access-vnqnv\") pod \"cert-manager-operator-controller-manager-57cd46d6d-fmhwf\" (UID: \"82ee4886-7435-48d7-bcbb-18b5c2f68758\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-fmhwf" Oct 02 10:02:33 crc kubenswrapper[4934]: I1002 10:02:33.973861 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-fmhwf" Oct 02 10:02:34 crc kubenswrapper[4934]: I1002 10:02:34.242756 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-fmhwf"] Oct 02 10:02:34 crc kubenswrapper[4934]: W1002 10:02:34.252218 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82ee4886_7435_48d7_bcbb_18b5c2f68758.slice/crio-197e7f303f30a0c0f386722823084faa740db59889490d280d6def234c653c2c WatchSource:0}: Error finding container 197e7f303f30a0c0f386722823084faa740db59889490d280d6def234c653c2c: Status 404 returned error can't find the container with id 197e7f303f30a0c0f386722823084faa740db59889490d280d6def234c653c2c Oct 02 10:02:34 crc kubenswrapper[4934]: I1002 10:02:34.428839 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-fmhwf" event={"ID":"82ee4886-7435-48d7-bcbb-18b5c2f68758","Type":"ContainerStarted","Data":"197e7f303f30a0c0f386722823084faa740db59889490d280d6def234c653c2c"} Oct 02 10:02:35 crc kubenswrapper[4934]: I1002 10:02:35.774881 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-jlmg7" Oct 02 10:02:41 crc kubenswrapper[4934]: I1002 10:02:41.474155 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-fmhwf" event={"ID":"82ee4886-7435-48d7-bcbb-18b5c2f68758","Type":"ContainerStarted","Data":"e371397f45a796779070359a733d1f689dc3be60d5e9730adbd6ab0cd5cea695"} Oct 02 10:02:41 crc kubenswrapper[4934]: I1002 10:02:41.491393 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-fmhwf" podStartSLOduration=1.864698325 podStartE2EDuration="8.491379256s" podCreationTimestamp="2025-10-02 10:02:33 +0000 UTC" firstStartedPulling="2025-10-02 10:02:34.254951728 +0000 UTC m=+826.007593250" lastFinishedPulling="2025-10-02 10:02:40.881632659 +0000 UTC m=+832.634274181" observedRunningTime="2025-10-02 10:02:41.489547406 +0000 UTC m=+833.242188928" watchObservedRunningTime="2025-10-02 10:02:41.491379256 +0000 UTC m=+833.244020778" Oct 02 10:02:44 crc kubenswrapper[4934]: I1002 10:02:44.252210 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-mfkpt"] Oct 02 10:02:44 crc kubenswrapper[4934]: I1002 10:02:44.252931 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-mfkpt" Oct 02 10:02:44 crc kubenswrapper[4934]: I1002 10:02:44.255277 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 02 10:02:44 crc kubenswrapper[4934]: I1002 10:02:44.255646 4934 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-4pvp7" Oct 02 10:02:44 crc kubenswrapper[4934]: I1002 10:02:44.255887 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 02 10:02:44 crc kubenswrapper[4934]: I1002 10:02:44.265268 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-mfkpt"] Oct 02 10:02:44 crc kubenswrapper[4934]: I1002 10:02:44.374287 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phwff\" (UniqueName: \"kubernetes.io/projected/473f4679-38f6-49b2-8f0b-d874ab84678e-kube-api-access-phwff\") pod \"cert-manager-webhook-d969966f-mfkpt\" (UID: \"473f4679-38f6-49b2-8f0b-d874ab84678e\") " pod="cert-manager/cert-manager-webhook-d969966f-mfkpt" Oct 02 10:02:44 crc kubenswrapper[4934]: I1002 10:02:44.374802 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/473f4679-38f6-49b2-8f0b-d874ab84678e-bound-sa-token\") pod \"cert-manager-webhook-d969966f-mfkpt\" (UID: \"473f4679-38f6-49b2-8f0b-d874ab84678e\") " pod="cert-manager/cert-manager-webhook-d969966f-mfkpt" Oct 02 10:02:44 crc kubenswrapper[4934]: I1002 10:02:44.476396 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phwff\" (UniqueName: \"kubernetes.io/projected/473f4679-38f6-49b2-8f0b-d874ab84678e-kube-api-access-phwff\") pod \"cert-manager-webhook-d969966f-mfkpt\" (UID: \"473f4679-38f6-49b2-8f0b-d874ab84678e\") " pod="cert-manager/cert-manager-webhook-d969966f-mfkpt" Oct 02 10:02:44 crc kubenswrapper[4934]: I1002 10:02:44.476452 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/473f4679-38f6-49b2-8f0b-d874ab84678e-bound-sa-token\") pod \"cert-manager-webhook-d969966f-mfkpt\" (UID: \"473f4679-38f6-49b2-8f0b-d874ab84678e\") " pod="cert-manager/cert-manager-webhook-d969966f-mfkpt" Oct 02 10:02:44 crc kubenswrapper[4934]: I1002 10:02:44.496304 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/473f4679-38f6-49b2-8f0b-d874ab84678e-bound-sa-token\") pod \"cert-manager-webhook-d969966f-mfkpt\" (UID: \"473f4679-38f6-49b2-8f0b-d874ab84678e\") " pod="cert-manager/cert-manager-webhook-d969966f-mfkpt" Oct 02 10:02:44 crc kubenswrapper[4934]: I1002 10:02:44.502736 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phwff\" (UniqueName: \"kubernetes.io/projected/473f4679-38f6-49b2-8f0b-d874ab84678e-kube-api-access-phwff\") pod \"cert-manager-webhook-d969966f-mfkpt\" (UID: \"473f4679-38f6-49b2-8f0b-d874ab84678e\") " pod="cert-manager/cert-manager-webhook-d969966f-mfkpt" Oct 02 10:02:44 crc kubenswrapper[4934]: I1002 10:02:44.572271 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-mfkpt" Oct 02 10:02:44 crc kubenswrapper[4934]: I1002 10:02:44.968326 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-mfkpt"] Oct 02 10:02:44 crc kubenswrapper[4934]: W1002 10:02:44.977467 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod473f4679_38f6_49b2_8f0b_d874ab84678e.slice/crio-6d0dc96f28b313374fd9bbfab9e61b08d3d05dabc05ce3177b8e89741079aae6 WatchSource:0}: Error finding container 6d0dc96f28b313374fd9bbfab9e61b08d3d05dabc05ce3177b8e89741079aae6: Status 404 returned error can't find the container with id 6d0dc96f28b313374fd9bbfab9e61b08d3d05dabc05ce3177b8e89741079aae6 Oct 02 10:02:45 crc kubenswrapper[4934]: I1002 10:02:45.495299 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-mfkpt" event={"ID":"473f4679-38f6-49b2-8f0b-d874ab84678e","Type":"ContainerStarted","Data":"6d0dc96f28b313374fd9bbfab9e61b08d3d05dabc05ce3177b8e89741079aae6"} Oct 02 10:02:47 crc kubenswrapper[4934]: I1002 10:02:47.389030 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-8w2z9"] Oct 02 10:02:47 crc kubenswrapper[4934]: I1002 10:02:47.390819 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-8w2z9" Oct 02 10:02:47 crc kubenswrapper[4934]: I1002 10:02:47.394792 4934 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-7bgwx" Oct 02 10:02:47 crc kubenswrapper[4934]: I1002 10:02:47.404397 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-8w2z9"] Oct 02 10:02:47 crc kubenswrapper[4934]: I1002 10:02:47.523353 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9fd4a62d-3068-4e62-8f90-7bf9c219ceee-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-8w2z9\" (UID: \"9fd4a62d-3068-4e62-8f90-7bf9c219ceee\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-8w2z9" Oct 02 10:02:47 crc kubenswrapper[4934]: I1002 10:02:47.523466 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dqhr\" (UniqueName: \"kubernetes.io/projected/9fd4a62d-3068-4e62-8f90-7bf9c219ceee-kube-api-access-7dqhr\") pod \"cert-manager-cainjector-7d9f95dbf-8w2z9\" (UID: \"9fd4a62d-3068-4e62-8f90-7bf9c219ceee\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-8w2z9" Oct 02 10:02:47 crc kubenswrapper[4934]: I1002 10:02:47.625188 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dqhr\" (UniqueName: \"kubernetes.io/projected/9fd4a62d-3068-4e62-8f90-7bf9c219ceee-kube-api-access-7dqhr\") pod \"cert-manager-cainjector-7d9f95dbf-8w2z9\" (UID: \"9fd4a62d-3068-4e62-8f90-7bf9c219ceee\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-8w2z9" Oct 02 10:02:47 crc kubenswrapper[4934]: I1002 10:02:47.625273 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9fd4a62d-3068-4e62-8f90-7bf9c219ceee-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-8w2z9\" (UID: \"9fd4a62d-3068-4e62-8f90-7bf9c219ceee\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-8w2z9" Oct 02 10:02:47 crc kubenswrapper[4934]: I1002 10:02:47.648752 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dqhr\" (UniqueName: \"kubernetes.io/projected/9fd4a62d-3068-4e62-8f90-7bf9c219ceee-kube-api-access-7dqhr\") pod \"cert-manager-cainjector-7d9f95dbf-8w2z9\" (UID: \"9fd4a62d-3068-4e62-8f90-7bf9c219ceee\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-8w2z9" Oct 02 10:02:47 crc kubenswrapper[4934]: I1002 10:02:47.666570 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9fd4a62d-3068-4e62-8f90-7bf9c219ceee-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-8w2z9\" (UID: \"9fd4a62d-3068-4e62-8f90-7bf9c219ceee\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-8w2z9" Oct 02 10:02:47 crc kubenswrapper[4934]: I1002 10:02:47.720539 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-8w2z9" Oct 02 10:02:49 crc kubenswrapper[4934]: I1002 10:02:49.321279 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-8w2z9"] Oct 02 10:02:49 crc kubenswrapper[4934]: W1002 10:02:49.331150 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9fd4a62d_3068_4e62_8f90_7bf9c219ceee.slice/crio-dcb67a94a4a9ce488aed40c5e0fa27a30291320bd135272cf4402a3d75693748 WatchSource:0}: Error finding container dcb67a94a4a9ce488aed40c5e0fa27a30291320bd135272cf4402a3d75693748: Status 404 returned error can't find the container with id dcb67a94a4a9ce488aed40c5e0fa27a30291320bd135272cf4402a3d75693748 Oct 02 10:02:49 crc kubenswrapper[4934]: I1002 10:02:49.528201 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-8w2z9" event={"ID":"9fd4a62d-3068-4e62-8f90-7bf9c219ceee","Type":"ContainerStarted","Data":"6f4097744cd5baf86e99f7fb435f1a0fe252e768416dccbd7e1faa29347b897a"} Oct 02 10:02:49 crc kubenswrapper[4934]: I1002 10:02:49.528630 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-8w2z9" event={"ID":"9fd4a62d-3068-4e62-8f90-7bf9c219ceee","Type":"ContainerStarted","Data":"dcb67a94a4a9ce488aed40c5e0fa27a30291320bd135272cf4402a3d75693748"} Oct 02 10:02:49 crc kubenswrapper[4934]: I1002 10:02:49.530531 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-mfkpt" event={"ID":"473f4679-38f6-49b2-8f0b-d874ab84678e","Type":"ContainerStarted","Data":"6f5174df73c667a9ecea204aad718fbd4e10c5f40567031f03c563ffe11ff9f5"} Oct 02 10:02:49 crc kubenswrapper[4934]: I1002 10:02:49.530721 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-d969966f-mfkpt" Oct 02 10:02:49 crc kubenswrapper[4934]: I1002 10:02:49.543283 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-8w2z9" podStartSLOduration=2.543266542 podStartE2EDuration="2.543266542s" podCreationTimestamp="2025-10-02 10:02:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:02:49.542290115 +0000 UTC m=+841.294931647" watchObservedRunningTime="2025-10-02 10:02:49.543266542 +0000 UTC m=+841.295908064" Oct 02 10:02:49 crc kubenswrapper[4934]: I1002 10:02:49.564273 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-d969966f-mfkpt" podStartSLOduration=1.361739993 podStartE2EDuration="5.564242411s" podCreationTimestamp="2025-10-02 10:02:44 +0000 UTC" firstStartedPulling="2025-10-02 10:02:44.979647812 +0000 UTC m=+836.732289334" lastFinishedPulling="2025-10-02 10:02:49.18215023 +0000 UTC m=+840.934791752" observedRunningTime="2025-10-02 10:02:49.559279264 +0000 UTC m=+841.311920786" watchObservedRunningTime="2025-10-02 10:02:49.564242411 +0000 UTC m=+841.316883973" Oct 02 10:02:54 crc kubenswrapper[4934]: I1002 10:02:54.574681 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-d969966f-mfkpt" Oct 02 10:03:03 crc kubenswrapper[4934]: I1002 10:03:03.442013 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lzbws"] Oct 02 10:03:03 crc kubenswrapper[4934]: I1002 10:03:03.444025 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lzbws" Oct 02 10:03:03 crc kubenswrapper[4934]: I1002 10:03:03.461442 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lzbws"] Oct 02 10:03:03 crc kubenswrapper[4934]: I1002 10:03:03.633767 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c67f1e6-4c3e-4854-94b4-17c78544db92-catalog-content\") pod \"certified-operators-lzbws\" (UID: \"5c67f1e6-4c3e-4854-94b4-17c78544db92\") " pod="openshift-marketplace/certified-operators-lzbws" Oct 02 10:03:03 crc kubenswrapper[4934]: I1002 10:03:03.633823 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6w8t\" (UniqueName: \"kubernetes.io/projected/5c67f1e6-4c3e-4854-94b4-17c78544db92-kube-api-access-w6w8t\") pod \"certified-operators-lzbws\" (UID: \"5c67f1e6-4c3e-4854-94b4-17c78544db92\") " pod="openshift-marketplace/certified-operators-lzbws" Oct 02 10:03:03 crc kubenswrapper[4934]: I1002 10:03:03.633871 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c67f1e6-4c3e-4854-94b4-17c78544db92-utilities\") pod \"certified-operators-lzbws\" (UID: \"5c67f1e6-4c3e-4854-94b4-17c78544db92\") " pod="openshift-marketplace/certified-operators-lzbws" Oct 02 10:03:03 crc kubenswrapper[4934]: I1002 10:03:03.735704 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c67f1e6-4c3e-4854-94b4-17c78544db92-catalog-content\") pod \"certified-operators-lzbws\" (UID: \"5c67f1e6-4c3e-4854-94b4-17c78544db92\") " pod="openshift-marketplace/certified-operators-lzbws" Oct 02 10:03:03 crc kubenswrapper[4934]: I1002 10:03:03.735764 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6w8t\" (UniqueName: \"kubernetes.io/projected/5c67f1e6-4c3e-4854-94b4-17c78544db92-kube-api-access-w6w8t\") pod \"certified-operators-lzbws\" (UID: \"5c67f1e6-4c3e-4854-94b4-17c78544db92\") " pod="openshift-marketplace/certified-operators-lzbws" Oct 02 10:03:03 crc kubenswrapper[4934]: I1002 10:03:03.735817 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c67f1e6-4c3e-4854-94b4-17c78544db92-utilities\") pod \"certified-operators-lzbws\" (UID: \"5c67f1e6-4c3e-4854-94b4-17c78544db92\") " pod="openshift-marketplace/certified-operators-lzbws" Oct 02 10:03:03 crc kubenswrapper[4934]: I1002 10:03:03.736590 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c67f1e6-4c3e-4854-94b4-17c78544db92-catalog-content\") pod \"certified-operators-lzbws\" (UID: \"5c67f1e6-4c3e-4854-94b4-17c78544db92\") " pod="openshift-marketplace/certified-operators-lzbws" Oct 02 10:03:03 crc kubenswrapper[4934]: I1002 10:03:03.736770 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c67f1e6-4c3e-4854-94b4-17c78544db92-utilities\") pod \"certified-operators-lzbws\" (UID: \"5c67f1e6-4c3e-4854-94b4-17c78544db92\") " pod="openshift-marketplace/certified-operators-lzbws" Oct 02 10:03:03 crc kubenswrapper[4934]: I1002 10:03:03.775211 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6w8t\" (UniqueName: \"kubernetes.io/projected/5c67f1e6-4c3e-4854-94b4-17c78544db92-kube-api-access-w6w8t\") pod \"certified-operators-lzbws\" (UID: \"5c67f1e6-4c3e-4854-94b4-17c78544db92\") " pod="openshift-marketplace/certified-operators-lzbws" Oct 02 10:03:04 crc kubenswrapper[4934]: I1002 10:03:04.073657 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lzbws" Oct 02 10:03:04 crc kubenswrapper[4934]: I1002 10:03:04.500675 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lzbws"] Oct 02 10:03:04 crc kubenswrapper[4934]: W1002 10:03:04.503502 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5c67f1e6_4c3e_4854_94b4_17c78544db92.slice/crio-e792cb01b6dd9b0fe6da9cfd4d3da21f3d3d73a0b71be0566744313c52eda976 WatchSource:0}: Error finding container e792cb01b6dd9b0fe6da9cfd4d3da21f3d3d73a0b71be0566744313c52eda976: Status 404 returned error can't find the container with id e792cb01b6dd9b0fe6da9cfd4d3da21f3d3d73a0b71be0566744313c52eda976 Oct 02 10:03:04 crc kubenswrapper[4934]: I1002 10:03:04.508046 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-gp6dr"] Oct 02 10:03:04 crc kubenswrapper[4934]: I1002 10:03:04.508999 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-gp6dr" Oct 02 10:03:04 crc kubenswrapper[4934]: I1002 10:03:04.511440 4934 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-gwbkh" Oct 02 10:03:04 crc kubenswrapper[4934]: I1002 10:03:04.515957 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-gp6dr"] Oct 02 10:03:04 crc kubenswrapper[4934]: I1002 10:03:04.554406 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhm7b\" (UniqueName: \"kubernetes.io/projected/30f8f452-7988-4522-9506-f6de0a8d17d3-kube-api-access-fhm7b\") pod \"cert-manager-7d4cc89fcb-gp6dr\" (UID: \"30f8f452-7988-4522-9506-f6de0a8d17d3\") " pod="cert-manager/cert-manager-7d4cc89fcb-gp6dr" Oct 02 10:03:04 crc kubenswrapper[4934]: I1002 10:03:04.554484 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/30f8f452-7988-4522-9506-f6de0a8d17d3-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-gp6dr\" (UID: \"30f8f452-7988-4522-9506-f6de0a8d17d3\") " pod="cert-manager/cert-manager-7d4cc89fcb-gp6dr" Oct 02 10:03:04 crc kubenswrapper[4934]: I1002 10:03:04.630865 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lzbws" event={"ID":"5c67f1e6-4c3e-4854-94b4-17c78544db92","Type":"ContainerStarted","Data":"e792cb01b6dd9b0fe6da9cfd4d3da21f3d3d73a0b71be0566744313c52eda976"} Oct 02 10:03:04 crc kubenswrapper[4934]: I1002 10:03:04.655056 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/30f8f452-7988-4522-9506-f6de0a8d17d3-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-gp6dr\" (UID: \"30f8f452-7988-4522-9506-f6de0a8d17d3\") " pod="cert-manager/cert-manager-7d4cc89fcb-gp6dr" Oct 02 10:03:04 crc kubenswrapper[4934]: I1002 10:03:04.655145 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhm7b\" (UniqueName: \"kubernetes.io/projected/30f8f452-7988-4522-9506-f6de0a8d17d3-kube-api-access-fhm7b\") pod \"cert-manager-7d4cc89fcb-gp6dr\" (UID: \"30f8f452-7988-4522-9506-f6de0a8d17d3\") " pod="cert-manager/cert-manager-7d4cc89fcb-gp6dr" Oct 02 10:03:04 crc kubenswrapper[4934]: I1002 10:03:04.671420 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/30f8f452-7988-4522-9506-f6de0a8d17d3-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-gp6dr\" (UID: \"30f8f452-7988-4522-9506-f6de0a8d17d3\") " pod="cert-manager/cert-manager-7d4cc89fcb-gp6dr" Oct 02 10:03:04 crc kubenswrapper[4934]: I1002 10:03:04.676547 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhm7b\" (UniqueName: \"kubernetes.io/projected/30f8f452-7988-4522-9506-f6de0a8d17d3-kube-api-access-fhm7b\") pod \"cert-manager-7d4cc89fcb-gp6dr\" (UID: \"30f8f452-7988-4522-9506-f6de0a8d17d3\") " pod="cert-manager/cert-manager-7d4cc89fcb-gp6dr" Oct 02 10:03:04 crc kubenswrapper[4934]: I1002 10:03:04.835762 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-gp6dr" Oct 02 10:03:05 crc kubenswrapper[4934]: I1002 10:03:05.243234 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-gp6dr"] Oct 02 10:03:05 crc kubenswrapper[4934]: W1002 10:03:05.281911 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30f8f452_7988_4522_9506_f6de0a8d17d3.slice/crio-6eaf0fe8d3ef2c2bf8dafeed811f8d8f56e93879af891e38053df569217cebf4 WatchSource:0}: Error finding container 6eaf0fe8d3ef2c2bf8dafeed811f8d8f56e93879af891e38053df569217cebf4: Status 404 returned error can't find the container with id 6eaf0fe8d3ef2c2bf8dafeed811f8d8f56e93879af891e38053df569217cebf4 Oct 02 10:03:05 crc kubenswrapper[4934]: I1002 10:03:05.638424 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-gp6dr" event={"ID":"30f8f452-7988-4522-9506-f6de0a8d17d3","Type":"ContainerStarted","Data":"1fc4a0518b01896ac9396f472ba97c9327f6d189462b881a9e6399c7b0cc6362"} Oct 02 10:03:05 crc kubenswrapper[4934]: I1002 10:03:05.638476 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-gp6dr" event={"ID":"30f8f452-7988-4522-9506-f6de0a8d17d3","Type":"ContainerStarted","Data":"6eaf0fe8d3ef2c2bf8dafeed811f8d8f56e93879af891e38053df569217cebf4"} Oct 02 10:03:05 crc kubenswrapper[4934]: I1002 10:03:05.640337 4934 generic.go:334] "Generic (PLEG): container finished" podID="5c67f1e6-4c3e-4854-94b4-17c78544db92" containerID="f5ad970268c1d50703d053b90c8834ba902b788ab9caceb4d3a6ba67c8b97764" exitCode=0 Oct 02 10:03:05 crc kubenswrapper[4934]: I1002 10:03:05.640374 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lzbws" event={"ID":"5c67f1e6-4c3e-4854-94b4-17c78544db92","Type":"ContainerDied","Data":"f5ad970268c1d50703d053b90c8834ba902b788ab9caceb4d3a6ba67c8b97764"} Oct 02 10:03:05 crc kubenswrapper[4934]: I1002 10:03:05.660965 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-7d4cc89fcb-gp6dr" podStartSLOduration=1.6609479249999999 podStartE2EDuration="1.660947925s" podCreationTimestamp="2025-10-02 10:03:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:03:05.658141578 +0000 UTC m=+857.410783100" watchObservedRunningTime="2025-10-02 10:03:05.660947925 +0000 UTC m=+857.413589447" Oct 02 10:03:07 crc kubenswrapper[4934]: I1002 10:03:07.653042 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lzbws" event={"ID":"5c67f1e6-4c3e-4854-94b4-17c78544db92","Type":"ContainerStarted","Data":"0e4739480489d5e7ef9af008f5c180f8691ed96c6fed88039c5437a7c000b602"} Oct 02 10:03:08 crc kubenswrapper[4934]: I1002 10:03:08.660293 4934 generic.go:334] "Generic (PLEG): container finished" podID="5c67f1e6-4c3e-4854-94b4-17c78544db92" containerID="0e4739480489d5e7ef9af008f5c180f8691ed96c6fed88039c5437a7c000b602" exitCode=0 Oct 02 10:03:08 crc kubenswrapper[4934]: I1002 10:03:08.660338 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lzbws" event={"ID":"5c67f1e6-4c3e-4854-94b4-17c78544db92","Type":"ContainerDied","Data":"0e4739480489d5e7ef9af008f5c180f8691ed96c6fed88039c5437a7c000b602"} Oct 02 10:03:09 crc kubenswrapper[4934]: I1002 10:03:09.667026 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lzbws" event={"ID":"5c67f1e6-4c3e-4854-94b4-17c78544db92","Type":"ContainerStarted","Data":"e7493072c6a465dd83c2671d4b4d0617ff8f767d7acfe9908000c7f93f672d2c"} Oct 02 10:03:09 crc kubenswrapper[4934]: I1002 10:03:09.686967 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lzbws" podStartSLOduration=3.145773383 podStartE2EDuration="6.686948139s" podCreationTimestamp="2025-10-02 10:03:03 +0000 UTC" firstStartedPulling="2025-10-02 10:03:05.643253057 +0000 UTC m=+857.395894579" lastFinishedPulling="2025-10-02 10:03:09.184427813 +0000 UTC m=+860.937069335" observedRunningTime="2025-10-02 10:03:09.685619093 +0000 UTC m=+861.438260615" watchObservedRunningTime="2025-10-02 10:03:09.686948139 +0000 UTC m=+861.439589661" Oct 02 10:03:11 crc kubenswrapper[4934]: I1002 10:03:11.012633 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-psh42"] Oct 02 10:03:11 crc kubenswrapper[4934]: I1002 10:03:11.013712 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-psh42" Oct 02 10:03:11 crc kubenswrapper[4934]: I1002 10:03:11.015908 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 02 10:03:11 crc kubenswrapper[4934]: I1002 10:03:11.016436 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-p27ng" Oct 02 10:03:11 crc kubenswrapper[4934]: I1002 10:03:11.016457 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 02 10:03:11 crc kubenswrapper[4934]: I1002 10:03:11.030666 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-psh42"] Oct 02 10:03:11 crc kubenswrapper[4934]: I1002 10:03:11.139918 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-js7jj\" (UniqueName: \"kubernetes.io/projected/cef5ed35-635f-4382-989b-2d48acf934d6-kube-api-access-js7jj\") pod \"openstack-operator-index-psh42\" (UID: \"cef5ed35-635f-4382-989b-2d48acf934d6\") " pod="openstack-operators/openstack-operator-index-psh42" Oct 02 10:03:11 crc kubenswrapper[4934]: I1002 10:03:11.240974 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-js7jj\" (UniqueName: \"kubernetes.io/projected/cef5ed35-635f-4382-989b-2d48acf934d6-kube-api-access-js7jj\") pod \"openstack-operator-index-psh42\" (UID: \"cef5ed35-635f-4382-989b-2d48acf934d6\") " pod="openstack-operators/openstack-operator-index-psh42" Oct 02 10:03:11 crc kubenswrapper[4934]: I1002 10:03:11.258360 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-js7jj\" (UniqueName: \"kubernetes.io/projected/cef5ed35-635f-4382-989b-2d48acf934d6-kube-api-access-js7jj\") pod \"openstack-operator-index-psh42\" (UID: \"cef5ed35-635f-4382-989b-2d48acf934d6\") " pod="openstack-operators/openstack-operator-index-psh42" Oct 02 10:03:11 crc kubenswrapper[4934]: I1002 10:03:11.332009 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-psh42" Oct 02 10:03:11 crc kubenswrapper[4934]: I1002 10:03:11.546561 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-psh42"] Oct 02 10:03:11 crc kubenswrapper[4934]: I1002 10:03:11.679785 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-psh42" event={"ID":"cef5ed35-635f-4382-989b-2d48acf934d6","Type":"ContainerStarted","Data":"181893b23599bd048c3873e7330d36f071e898346ae653234ff5fbdf44020f5f"} Oct 02 10:03:13 crc kubenswrapper[4934]: I1002 10:03:13.696511 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-psh42" event={"ID":"cef5ed35-635f-4382-989b-2d48acf934d6","Type":"ContainerStarted","Data":"24f60fb2e8ab7b5210c34cac39247c9c6b82e460c5a963c6c2c8ab23cf65c0e2"} Oct 02 10:03:13 crc kubenswrapper[4934]: I1002 10:03:13.718845 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-psh42" podStartSLOduration=1.405410177 podStartE2EDuration="2.718821976s" podCreationTimestamp="2025-10-02 10:03:11 +0000 UTC" firstStartedPulling="2025-10-02 10:03:11.555028554 +0000 UTC m=+863.307670086" lastFinishedPulling="2025-10-02 10:03:12.868440353 +0000 UTC m=+864.621081885" observedRunningTime="2025-10-02 10:03:13.717269463 +0000 UTC m=+865.469911005" watchObservedRunningTime="2025-10-02 10:03:13.718821976 +0000 UTC m=+865.471463518" Oct 02 10:03:14 crc kubenswrapper[4934]: I1002 10:03:14.074068 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lzbws" Oct 02 10:03:14 crc kubenswrapper[4934]: I1002 10:03:14.074132 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lzbws" Oct 02 10:03:14 crc kubenswrapper[4934]: I1002 10:03:14.116862 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lzbws" Oct 02 10:03:14 crc kubenswrapper[4934]: I1002 10:03:14.740539 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lzbws" Oct 02 10:03:16 crc kubenswrapper[4934]: I1002 10:03:16.407643 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-psh42"] Oct 02 10:03:16 crc kubenswrapper[4934]: I1002 10:03:16.408282 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-psh42" podUID="cef5ed35-635f-4382-989b-2d48acf934d6" containerName="registry-server" containerID="cri-o://24f60fb2e8ab7b5210c34cac39247c9c6b82e460c5a963c6c2c8ab23cf65c0e2" gracePeriod=2 Oct 02 10:03:16 crc kubenswrapper[4934]: I1002 10:03:16.723835 4934 generic.go:334] "Generic (PLEG): container finished" podID="cef5ed35-635f-4382-989b-2d48acf934d6" containerID="24f60fb2e8ab7b5210c34cac39247c9c6b82e460c5a963c6c2c8ab23cf65c0e2" exitCode=0 Oct 02 10:03:16 crc kubenswrapper[4934]: I1002 10:03:16.723862 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-psh42" event={"ID":"cef5ed35-635f-4382-989b-2d48acf934d6","Type":"ContainerDied","Data":"24f60fb2e8ab7b5210c34cac39247c9c6b82e460c5a963c6c2c8ab23cf65c0e2"} Oct 02 10:03:16 crc kubenswrapper[4934]: I1002 10:03:16.787369 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-psh42" Oct 02 10:03:16 crc kubenswrapper[4934]: I1002 10:03:16.816514 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-js7jj\" (UniqueName: \"kubernetes.io/projected/cef5ed35-635f-4382-989b-2d48acf934d6-kube-api-access-js7jj\") pod \"cef5ed35-635f-4382-989b-2d48acf934d6\" (UID: \"cef5ed35-635f-4382-989b-2d48acf934d6\") " Oct 02 10:03:16 crc kubenswrapper[4934]: I1002 10:03:16.822530 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cef5ed35-635f-4382-989b-2d48acf934d6-kube-api-access-js7jj" (OuterVolumeSpecName: "kube-api-access-js7jj") pod "cef5ed35-635f-4382-989b-2d48acf934d6" (UID: "cef5ed35-635f-4382-989b-2d48acf934d6"). InnerVolumeSpecName "kube-api-access-js7jj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:03:16 crc kubenswrapper[4934]: I1002 10:03:16.917966 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-js7jj\" (UniqueName: \"kubernetes.io/projected/cef5ed35-635f-4382-989b-2d48acf934d6-kube-api-access-js7jj\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:17 crc kubenswrapper[4934]: I1002 10:03:17.010248 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-fvkxx"] Oct 02 10:03:17 crc kubenswrapper[4934]: E1002 10:03:17.010510 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cef5ed35-635f-4382-989b-2d48acf934d6" containerName="registry-server" Oct 02 10:03:17 crc kubenswrapper[4934]: I1002 10:03:17.010525 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="cef5ed35-635f-4382-989b-2d48acf934d6" containerName="registry-server" Oct 02 10:03:17 crc kubenswrapper[4934]: I1002 10:03:17.010740 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="cef5ed35-635f-4382-989b-2d48acf934d6" containerName="registry-server" Oct 02 10:03:17 crc kubenswrapper[4934]: I1002 10:03:17.011207 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fvkxx" Oct 02 10:03:17 crc kubenswrapper[4934]: I1002 10:03:17.032704 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dn8x9\" (UniqueName: \"kubernetes.io/projected/c66499be-f2ac-4f3e-82e3-e90eed8922ce-kube-api-access-dn8x9\") pod \"openstack-operator-index-fvkxx\" (UID: \"c66499be-f2ac-4f3e-82e3-e90eed8922ce\") " pod="openstack-operators/openstack-operator-index-fvkxx" Oct 02 10:03:17 crc kubenswrapper[4934]: I1002 10:03:17.038059 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-fvkxx"] Oct 02 10:03:17 crc kubenswrapper[4934]: I1002 10:03:17.133912 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dn8x9\" (UniqueName: \"kubernetes.io/projected/c66499be-f2ac-4f3e-82e3-e90eed8922ce-kube-api-access-dn8x9\") pod \"openstack-operator-index-fvkxx\" (UID: \"c66499be-f2ac-4f3e-82e3-e90eed8922ce\") " pod="openstack-operators/openstack-operator-index-fvkxx" Oct 02 10:03:17 crc kubenswrapper[4934]: I1002 10:03:17.152606 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dn8x9\" (UniqueName: \"kubernetes.io/projected/c66499be-f2ac-4f3e-82e3-e90eed8922ce-kube-api-access-dn8x9\") pod \"openstack-operator-index-fvkxx\" (UID: \"c66499be-f2ac-4f3e-82e3-e90eed8922ce\") " pod="openstack-operators/openstack-operator-index-fvkxx" Oct 02 10:03:17 crc kubenswrapper[4934]: I1002 10:03:17.344611 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fvkxx" Oct 02 10:03:17 crc kubenswrapper[4934]: I1002 10:03:17.522492 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-fvkxx"] Oct 02 10:03:17 crc kubenswrapper[4934]: W1002 10:03:17.529200 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc66499be_f2ac_4f3e_82e3_e90eed8922ce.slice/crio-46dd00e6065f419ed06093b54935db1f4c224a3605d76df7e86ba4e9e285e826 WatchSource:0}: Error finding container 46dd00e6065f419ed06093b54935db1f4c224a3605d76df7e86ba4e9e285e826: Status 404 returned error can't find the container with id 46dd00e6065f419ed06093b54935db1f4c224a3605d76df7e86ba4e9e285e826 Oct 02 10:03:17 crc kubenswrapper[4934]: I1002 10:03:17.732561 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-psh42" event={"ID":"cef5ed35-635f-4382-989b-2d48acf934d6","Type":"ContainerDied","Data":"181893b23599bd048c3873e7330d36f071e898346ae653234ff5fbdf44020f5f"} Oct 02 10:03:17 crc kubenswrapper[4934]: I1002 10:03:17.732899 4934 scope.go:117] "RemoveContainer" containerID="24f60fb2e8ab7b5210c34cac39247c9c6b82e460c5a963c6c2c8ab23cf65c0e2" Oct 02 10:03:17 crc kubenswrapper[4934]: I1002 10:03:17.732665 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-psh42" Oct 02 10:03:17 crc kubenswrapper[4934]: I1002 10:03:17.735395 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fvkxx" event={"ID":"c66499be-f2ac-4f3e-82e3-e90eed8922ce","Type":"ContainerStarted","Data":"46dd00e6065f419ed06093b54935db1f4c224a3605d76df7e86ba4e9e285e826"} Oct 02 10:03:17 crc kubenswrapper[4934]: I1002 10:03:17.761365 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-psh42"] Oct 02 10:03:17 crc kubenswrapper[4934]: I1002 10:03:17.768828 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-psh42"] Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.005419 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lzbws"] Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.005679 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lzbws" podUID="5c67f1e6-4c3e-4854-94b4-17c78544db92" containerName="registry-server" containerID="cri-o://e7493072c6a465dd83c2671d4b4d0617ff8f767d7acfe9908000c7f93f672d2c" gracePeriod=2 Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.447418 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lzbws" Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.556874 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6w8t\" (UniqueName: \"kubernetes.io/projected/5c67f1e6-4c3e-4854-94b4-17c78544db92-kube-api-access-w6w8t\") pod \"5c67f1e6-4c3e-4854-94b4-17c78544db92\" (UID: \"5c67f1e6-4c3e-4854-94b4-17c78544db92\") " Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.558078 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c67f1e6-4c3e-4854-94b4-17c78544db92-catalog-content\") pod \"5c67f1e6-4c3e-4854-94b4-17c78544db92\" (UID: \"5c67f1e6-4c3e-4854-94b4-17c78544db92\") " Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.558269 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c67f1e6-4c3e-4854-94b4-17c78544db92-utilities\") pod \"5c67f1e6-4c3e-4854-94b4-17c78544db92\" (UID: \"5c67f1e6-4c3e-4854-94b4-17c78544db92\") " Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.560030 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c67f1e6-4c3e-4854-94b4-17c78544db92-utilities" (OuterVolumeSpecName: "utilities") pod "5c67f1e6-4c3e-4854-94b4-17c78544db92" (UID: "5c67f1e6-4c3e-4854-94b4-17c78544db92"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.564097 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c67f1e6-4c3e-4854-94b4-17c78544db92-kube-api-access-w6w8t" (OuterVolumeSpecName: "kube-api-access-w6w8t") pod "5c67f1e6-4c3e-4854-94b4-17c78544db92" (UID: "5c67f1e6-4c3e-4854-94b4-17c78544db92"). InnerVolumeSpecName "kube-api-access-w6w8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.605917 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c67f1e6-4c3e-4854-94b4-17c78544db92-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5c67f1e6-4c3e-4854-94b4-17c78544db92" (UID: "5c67f1e6-4c3e-4854-94b4-17c78544db92"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.659945 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c67f1e6-4c3e-4854-94b4-17c78544db92-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.660005 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6w8t\" (UniqueName: \"kubernetes.io/projected/5c67f1e6-4c3e-4854-94b4-17c78544db92-kube-api-access-w6w8t\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.660019 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c67f1e6-4c3e-4854-94b4-17c78544db92-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.749981 4934 generic.go:334] "Generic (PLEG): container finished" podID="5c67f1e6-4c3e-4854-94b4-17c78544db92" containerID="e7493072c6a465dd83c2671d4b4d0617ff8f767d7acfe9908000c7f93f672d2c" exitCode=0 Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.750072 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lzbws" Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.750060 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lzbws" event={"ID":"5c67f1e6-4c3e-4854-94b4-17c78544db92","Type":"ContainerDied","Data":"e7493072c6a465dd83c2671d4b4d0617ff8f767d7acfe9908000c7f93f672d2c"} Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.750673 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lzbws" event={"ID":"5c67f1e6-4c3e-4854-94b4-17c78544db92","Type":"ContainerDied","Data":"e792cb01b6dd9b0fe6da9cfd4d3da21f3d3d73a0b71be0566744313c52eda976"} Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.750725 4934 scope.go:117] "RemoveContainer" containerID="e7493072c6a465dd83c2671d4b4d0617ff8f767d7acfe9908000c7f93f672d2c" Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.752828 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fvkxx" event={"ID":"c66499be-f2ac-4f3e-82e3-e90eed8922ce","Type":"ContainerStarted","Data":"caea0f3e7f1f403f91c95f62ca180633f2010474b4224e19017e0fe530b336af"} Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.780469 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-fvkxx" podStartSLOduration=2.136051523 podStartE2EDuration="2.780449838s" podCreationTimestamp="2025-10-02 10:03:16 +0000 UTC" firstStartedPulling="2025-10-02 10:03:17.534997496 +0000 UTC m=+869.287639018" lastFinishedPulling="2025-10-02 10:03:18.179395821 +0000 UTC m=+869.932037333" observedRunningTime="2025-10-02 10:03:18.775410379 +0000 UTC m=+870.528051911" watchObservedRunningTime="2025-10-02 10:03:18.780449838 +0000 UTC m=+870.533091360" Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.786125 4934 scope.go:117] "RemoveContainer" containerID="0e4739480489d5e7ef9af008f5c180f8691ed96c6fed88039c5437a7c000b602" Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.807567 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lzbws"] Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.815887 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lzbws"] Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.828843 4934 scope.go:117] "RemoveContainer" containerID="f5ad970268c1d50703d053b90c8834ba902b788ab9caceb4d3a6ba67c8b97764" Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.845706 4934 scope.go:117] "RemoveContainer" containerID="e7493072c6a465dd83c2671d4b4d0617ff8f767d7acfe9908000c7f93f672d2c" Oct 02 10:03:18 crc kubenswrapper[4934]: E1002 10:03:18.846053 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7493072c6a465dd83c2671d4b4d0617ff8f767d7acfe9908000c7f93f672d2c\": container with ID starting with e7493072c6a465dd83c2671d4b4d0617ff8f767d7acfe9908000c7f93f672d2c not found: ID does not exist" containerID="e7493072c6a465dd83c2671d4b4d0617ff8f767d7acfe9908000c7f93f672d2c" Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.846085 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7493072c6a465dd83c2671d4b4d0617ff8f767d7acfe9908000c7f93f672d2c"} err="failed to get container status \"e7493072c6a465dd83c2671d4b4d0617ff8f767d7acfe9908000c7f93f672d2c\": rpc error: code = NotFound desc = could not find container \"e7493072c6a465dd83c2671d4b4d0617ff8f767d7acfe9908000c7f93f672d2c\": container with ID starting with e7493072c6a465dd83c2671d4b4d0617ff8f767d7acfe9908000c7f93f672d2c not found: ID does not exist" Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.846106 4934 scope.go:117] "RemoveContainer" containerID="0e4739480489d5e7ef9af008f5c180f8691ed96c6fed88039c5437a7c000b602" Oct 02 10:03:18 crc kubenswrapper[4934]: E1002 10:03:18.846541 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e4739480489d5e7ef9af008f5c180f8691ed96c6fed88039c5437a7c000b602\": container with ID starting with 0e4739480489d5e7ef9af008f5c180f8691ed96c6fed88039c5437a7c000b602 not found: ID does not exist" containerID="0e4739480489d5e7ef9af008f5c180f8691ed96c6fed88039c5437a7c000b602" Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.846564 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e4739480489d5e7ef9af008f5c180f8691ed96c6fed88039c5437a7c000b602"} err="failed to get container status \"0e4739480489d5e7ef9af008f5c180f8691ed96c6fed88039c5437a7c000b602\": rpc error: code = NotFound desc = could not find container \"0e4739480489d5e7ef9af008f5c180f8691ed96c6fed88039c5437a7c000b602\": container with ID starting with 0e4739480489d5e7ef9af008f5c180f8691ed96c6fed88039c5437a7c000b602 not found: ID does not exist" Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.846619 4934 scope.go:117] "RemoveContainer" containerID="f5ad970268c1d50703d053b90c8834ba902b788ab9caceb4d3a6ba67c8b97764" Oct 02 10:03:18 crc kubenswrapper[4934]: E1002 10:03:18.846809 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5ad970268c1d50703d053b90c8834ba902b788ab9caceb4d3a6ba67c8b97764\": container with ID starting with f5ad970268c1d50703d053b90c8834ba902b788ab9caceb4d3a6ba67c8b97764 not found: ID does not exist" containerID="f5ad970268c1d50703d053b90c8834ba902b788ab9caceb4d3a6ba67c8b97764" Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.846839 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5ad970268c1d50703d053b90c8834ba902b788ab9caceb4d3a6ba67c8b97764"} err="failed to get container status \"f5ad970268c1d50703d053b90c8834ba902b788ab9caceb4d3a6ba67c8b97764\": rpc error: code = NotFound desc = could not find container \"f5ad970268c1d50703d053b90c8834ba902b788ab9caceb4d3a6ba67c8b97764\": container with ID starting with f5ad970268c1d50703d053b90c8834ba902b788ab9caceb4d3a6ba67c8b97764 not found: ID does not exist" Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.922604 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c67f1e6-4c3e-4854-94b4-17c78544db92" path="/var/lib/kubelet/pods/5c67f1e6-4c3e-4854-94b4-17c78544db92/volumes" Oct 02 10:03:18 crc kubenswrapper[4934]: I1002 10:03:18.923297 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cef5ed35-635f-4382-989b-2d48acf934d6" path="/var/lib/kubelet/pods/cef5ed35-635f-4382-989b-2d48acf934d6/volumes" Oct 02 10:03:27 crc kubenswrapper[4934]: I1002 10:03:27.345277 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-fvkxx" Oct 02 10:03:27 crc kubenswrapper[4934]: I1002 10:03:27.345819 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-fvkxx" Oct 02 10:03:27 crc kubenswrapper[4934]: I1002 10:03:27.390340 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-fvkxx" Oct 02 10:03:27 crc kubenswrapper[4934]: I1002 10:03:27.840253 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-fvkxx" Oct 02 10:03:38 crc kubenswrapper[4934]: I1002 10:03:38.440072 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:03:38 crc kubenswrapper[4934]: I1002 10:03:38.440518 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:03:41 crc kubenswrapper[4934]: I1002 10:03:41.264305 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5"] Oct 02 10:03:41 crc kubenswrapper[4934]: E1002 10:03:41.264721 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c67f1e6-4c3e-4854-94b4-17c78544db92" containerName="registry-server" Oct 02 10:03:41 crc kubenswrapper[4934]: I1002 10:03:41.264743 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c67f1e6-4c3e-4854-94b4-17c78544db92" containerName="registry-server" Oct 02 10:03:41 crc kubenswrapper[4934]: E1002 10:03:41.264776 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c67f1e6-4c3e-4854-94b4-17c78544db92" containerName="extract-utilities" Oct 02 10:03:41 crc kubenswrapper[4934]: I1002 10:03:41.264787 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c67f1e6-4c3e-4854-94b4-17c78544db92" containerName="extract-utilities" Oct 02 10:03:41 crc kubenswrapper[4934]: E1002 10:03:41.264814 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c67f1e6-4c3e-4854-94b4-17c78544db92" containerName="extract-content" Oct 02 10:03:41 crc kubenswrapper[4934]: I1002 10:03:41.264826 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c67f1e6-4c3e-4854-94b4-17c78544db92" containerName="extract-content" Oct 02 10:03:41 crc kubenswrapper[4934]: I1002 10:03:41.264986 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c67f1e6-4c3e-4854-94b4-17c78544db92" containerName="registry-server" Oct 02 10:03:41 crc kubenswrapper[4934]: I1002 10:03:41.266370 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5" Oct 02 10:03:41 crc kubenswrapper[4934]: I1002 10:03:41.270801 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-cx6cf" Oct 02 10:03:41 crc kubenswrapper[4934]: I1002 10:03:41.279104 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5"] Oct 02 10:03:41 crc kubenswrapper[4934]: I1002 10:03:41.365039 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jp2pm\" (UniqueName: \"kubernetes.io/projected/94f3473a-92cb-428a-9cec-ad4de77d6a66-kube-api-access-jp2pm\") pod \"8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5\" (UID: \"94f3473a-92cb-428a-9cec-ad4de77d6a66\") " pod="openstack-operators/8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5" Oct 02 10:03:41 crc kubenswrapper[4934]: I1002 10:03:41.365439 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/94f3473a-92cb-428a-9cec-ad4de77d6a66-bundle\") pod \"8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5\" (UID: \"94f3473a-92cb-428a-9cec-ad4de77d6a66\") " pod="openstack-operators/8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5" Oct 02 10:03:41 crc kubenswrapper[4934]: I1002 10:03:41.365497 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/94f3473a-92cb-428a-9cec-ad4de77d6a66-util\") pod \"8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5\" (UID: \"94f3473a-92cb-428a-9cec-ad4de77d6a66\") " pod="openstack-operators/8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5" Oct 02 10:03:41 crc kubenswrapper[4934]: I1002 10:03:41.467242 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/94f3473a-92cb-428a-9cec-ad4de77d6a66-util\") pod \"8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5\" (UID: \"94f3473a-92cb-428a-9cec-ad4de77d6a66\") " pod="openstack-operators/8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5" Oct 02 10:03:41 crc kubenswrapper[4934]: I1002 10:03:41.467421 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jp2pm\" (UniqueName: \"kubernetes.io/projected/94f3473a-92cb-428a-9cec-ad4de77d6a66-kube-api-access-jp2pm\") pod \"8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5\" (UID: \"94f3473a-92cb-428a-9cec-ad4de77d6a66\") " pod="openstack-operators/8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5" Oct 02 10:03:41 crc kubenswrapper[4934]: I1002 10:03:41.467501 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/94f3473a-92cb-428a-9cec-ad4de77d6a66-bundle\") pod \"8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5\" (UID: \"94f3473a-92cb-428a-9cec-ad4de77d6a66\") " pod="openstack-operators/8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5" Oct 02 10:03:41 crc kubenswrapper[4934]: I1002 10:03:41.467824 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/94f3473a-92cb-428a-9cec-ad4de77d6a66-util\") pod \"8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5\" (UID: \"94f3473a-92cb-428a-9cec-ad4de77d6a66\") " pod="openstack-operators/8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5" Oct 02 10:03:41 crc kubenswrapper[4934]: I1002 10:03:41.468035 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/94f3473a-92cb-428a-9cec-ad4de77d6a66-bundle\") pod \"8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5\" (UID: \"94f3473a-92cb-428a-9cec-ad4de77d6a66\") " pod="openstack-operators/8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5" Oct 02 10:03:41 crc kubenswrapper[4934]: I1002 10:03:41.496681 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jp2pm\" (UniqueName: \"kubernetes.io/projected/94f3473a-92cb-428a-9cec-ad4de77d6a66-kube-api-access-jp2pm\") pod \"8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5\" (UID: \"94f3473a-92cb-428a-9cec-ad4de77d6a66\") " pod="openstack-operators/8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5" Oct 02 10:03:41 crc kubenswrapper[4934]: I1002 10:03:41.584989 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5" Oct 02 10:03:42 crc kubenswrapper[4934]: I1002 10:03:42.050992 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5"] Oct 02 10:03:42 crc kubenswrapper[4934]: I1002 10:03:42.924740 4934 generic.go:334] "Generic (PLEG): container finished" podID="94f3473a-92cb-428a-9cec-ad4de77d6a66" containerID="a8bd40dfd4bd6292e736c1bef7148ae3e783801c6a1327ebf040aeba6dcf83d6" exitCode=0 Oct 02 10:03:42 crc kubenswrapper[4934]: I1002 10:03:42.924824 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5" event={"ID":"94f3473a-92cb-428a-9cec-ad4de77d6a66","Type":"ContainerDied","Data":"a8bd40dfd4bd6292e736c1bef7148ae3e783801c6a1327ebf040aeba6dcf83d6"} Oct 02 10:03:42 crc kubenswrapper[4934]: I1002 10:03:42.925098 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5" event={"ID":"94f3473a-92cb-428a-9cec-ad4de77d6a66","Type":"ContainerStarted","Data":"a6aedad806755bca54284d0e7a8e1208eef2cdba0434c26e92fd84edfe8b6eab"} Oct 02 10:03:43 crc kubenswrapper[4934]: I1002 10:03:43.937184 4934 generic.go:334] "Generic (PLEG): container finished" podID="94f3473a-92cb-428a-9cec-ad4de77d6a66" containerID="f455dd2c13c88784bcadbd005af19adbd070ec150147627745e1a2fdb4680f8d" exitCode=0 Oct 02 10:03:43 crc kubenswrapper[4934]: I1002 10:03:43.937262 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5" event={"ID":"94f3473a-92cb-428a-9cec-ad4de77d6a66","Type":"ContainerDied","Data":"f455dd2c13c88784bcadbd005af19adbd070ec150147627745e1a2fdb4680f8d"} Oct 02 10:03:44 crc kubenswrapper[4934]: I1002 10:03:44.946156 4934 generic.go:334] "Generic (PLEG): container finished" podID="94f3473a-92cb-428a-9cec-ad4de77d6a66" containerID="c0ef1de14588657d3ee32689705a80404a07c2ddd00d23f9073796f016ef6fcf" exitCode=0 Oct 02 10:03:44 crc kubenswrapper[4934]: I1002 10:03:44.947708 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5" event={"ID":"94f3473a-92cb-428a-9cec-ad4de77d6a66","Type":"ContainerDied","Data":"c0ef1de14588657d3ee32689705a80404a07c2ddd00d23f9073796f016ef6fcf"} Oct 02 10:03:46 crc kubenswrapper[4934]: I1002 10:03:46.191251 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5" Oct 02 10:03:46 crc kubenswrapper[4934]: I1002 10:03:46.338130 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/94f3473a-92cb-428a-9cec-ad4de77d6a66-util\") pod \"94f3473a-92cb-428a-9cec-ad4de77d6a66\" (UID: \"94f3473a-92cb-428a-9cec-ad4de77d6a66\") " Oct 02 10:03:46 crc kubenswrapper[4934]: I1002 10:03:46.338545 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jp2pm\" (UniqueName: \"kubernetes.io/projected/94f3473a-92cb-428a-9cec-ad4de77d6a66-kube-api-access-jp2pm\") pod \"94f3473a-92cb-428a-9cec-ad4de77d6a66\" (UID: \"94f3473a-92cb-428a-9cec-ad4de77d6a66\") " Oct 02 10:03:46 crc kubenswrapper[4934]: I1002 10:03:46.338985 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/94f3473a-92cb-428a-9cec-ad4de77d6a66-bundle\") pod \"94f3473a-92cb-428a-9cec-ad4de77d6a66\" (UID: \"94f3473a-92cb-428a-9cec-ad4de77d6a66\") " Oct 02 10:03:46 crc kubenswrapper[4934]: I1002 10:03:46.341462 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94f3473a-92cb-428a-9cec-ad4de77d6a66-bundle" (OuterVolumeSpecName: "bundle") pod "94f3473a-92cb-428a-9cec-ad4de77d6a66" (UID: "94f3473a-92cb-428a-9cec-ad4de77d6a66"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:03:46 crc kubenswrapper[4934]: I1002 10:03:46.350025 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94f3473a-92cb-428a-9cec-ad4de77d6a66-kube-api-access-jp2pm" (OuterVolumeSpecName: "kube-api-access-jp2pm") pod "94f3473a-92cb-428a-9cec-ad4de77d6a66" (UID: "94f3473a-92cb-428a-9cec-ad4de77d6a66"). InnerVolumeSpecName "kube-api-access-jp2pm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:03:46 crc kubenswrapper[4934]: I1002 10:03:46.361152 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94f3473a-92cb-428a-9cec-ad4de77d6a66-util" (OuterVolumeSpecName: "util") pod "94f3473a-92cb-428a-9cec-ad4de77d6a66" (UID: "94f3473a-92cb-428a-9cec-ad4de77d6a66"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:03:46 crc kubenswrapper[4934]: I1002 10:03:46.441672 4934 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/94f3473a-92cb-428a-9cec-ad4de77d6a66-util\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:46 crc kubenswrapper[4934]: I1002 10:03:46.441710 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jp2pm\" (UniqueName: \"kubernetes.io/projected/94f3473a-92cb-428a-9cec-ad4de77d6a66-kube-api-access-jp2pm\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:46 crc kubenswrapper[4934]: I1002 10:03:46.441726 4934 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/94f3473a-92cb-428a-9cec-ad4de77d6a66-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:46 crc kubenswrapper[4934]: I1002 10:03:46.968363 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5" event={"ID":"94f3473a-92cb-428a-9cec-ad4de77d6a66","Type":"ContainerDied","Data":"a6aedad806755bca54284d0e7a8e1208eef2cdba0434c26e92fd84edfe8b6eab"} Oct 02 10:03:46 crc kubenswrapper[4934]: I1002 10:03:46.968423 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6aedad806755bca54284d0e7a8e1208eef2cdba0434c26e92fd84edfe8b6eab" Oct 02 10:03:46 crc kubenswrapper[4934]: I1002 10:03:46.968489 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5" Oct 02 10:03:50 crc kubenswrapper[4934]: I1002 10:03:50.242755 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7c7c8d99ff-g2x7f"] Oct 02 10:03:50 crc kubenswrapper[4934]: E1002 10:03:50.243504 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94f3473a-92cb-428a-9cec-ad4de77d6a66" containerName="util" Oct 02 10:03:50 crc kubenswrapper[4934]: I1002 10:03:50.243520 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="94f3473a-92cb-428a-9cec-ad4de77d6a66" containerName="util" Oct 02 10:03:50 crc kubenswrapper[4934]: E1002 10:03:50.243537 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94f3473a-92cb-428a-9cec-ad4de77d6a66" containerName="extract" Oct 02 10:03:50 crc kubenswrapper[4934]: I1002 10:03:50.243544 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="94f3473a-92cb-428a-9cec-ad4de77d6a66" containerName="extract" Oct 02 10:03:50 crc kubenswrapper[4934]: E1002 10:03:50.243555 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94f3473a-92cb-428a-9cec-ad4de77d6a66" containerName="pull" Oct 02 10:03:50 crc kubenswrapper[4934]: I1002 10:03:50.243561 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="94f3473a-92cb-428a-9cec-ad4de77d6a66" containerName="pull" Oct 02 10:03:50 crc kubenswrapper[4934]: I1002 10:03:50.243686 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="94f3473a-92cb-428a-9cec-ad4de77d6a66" containerName="extract" Oct 02 10:03:50 crc kubenswrapper[4934]: I1002 10:03:50.244283 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7c7c8d99ff-g2x7f" Oct 02 10:03:50 crc kubenswrapper[4934]: I1002 10:03:50.245935 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-sp6fn" Oct 02 10:03:50 crc kubenswrapper[4934]: I1002 10:03:50.289136 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7c7c8d99ff-g2x7f"] Oct 02 10:03:50 crc kubenswrapper[4934]: I1002 10:03:50.391290 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkfl2\" (UniqueName: \"kubernetes.io/projected/5d42fe89-99f8-436c-8401-835532319e14-kube-api-access-fkfl2\") pod \"openstack-operator-controller-operator-7c7c8d99ff-g2x7f\" (UID: \"5d42fe89-99f8-436c-8401-835532319e14\") " pod="openstack-operators/openstack-operator-controller-operator-7c7c8d99ff-g2x7f" Oct 02 10:03:50 crc kubenswrapper[4934]: I1002 10:03:50.493174 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkfl2\" (UniqueName: \"kubernetes.io/projected/5d42fe89-99f8-436c-8401-835532319e14-kube-api-access-fkfl2\") pod \"openstack-operator-controller-operator-7c7c8d99ff-g2x7f\" (UID: \"5d42fe89-99f8-436c-8401-835532319e14\") " pod="openstack-operators/openstack-operator-controller-operator-7c7c8d99ff-g2x7f" Oct 02 10:03:50 crc kubenswrapper[4934]: I1002 10:03:50.510771 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkfl2\" (UniqueName: \"kubernetes.io/projected/5d42fe89-99f8-436c-8401-835532319e14-kube-api-access-fkfl2\") pod \"openstack-operator-controller-operator-7c7c8d99ff-g2x7f\" (UID: \"5d42fe89-99f8-436c-8401-835532319e14\") " pod="openstack-operators/openstack-operator-controller-operator-7c7c8d99ff-g2x7f" Oct 02 10:03:50 crc kubenswrapper[4934]: I1002 10:03:50.560755 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7c7c8d99ff-g2x7f" Oct 02 10:03:50 crc kubenswrapper[4934]: I1002 10:03:50.810167 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7c7c8d99ff-g2x7f"] Oct 02 10:03:50 crc kubenswrapper[4934]: I1002 10:03:50.993765 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7c7c8d99ff-g2x7f" event={"ID":"5d42fe89-99f8-436c-8401-835532319e14","Type":"ContainerStarted","Data":"f939da1075f4e762efb0938a6dcc01300e0d5bfc348aa2e0eb2ac2c4befeb5d5"} Oct 02 10:03:56 crc kubenswrapper[4934]: I1002 10:03:56.027287 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7c7c8d99ff-g2x7f" event={"ID":"5d42fe89-99f8-436c-8401-835532319e14","Type":"ContainerStarted","Data":"46de7172398932737e6c464be2244e14d3116368a506efae8f4ec0979f12419d"} Oct 02 10:03:58 crc kubenswrapper[4934]: I1002 10:03:58.040196 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7c7c8d99ff-g2x7f" event={"ID":"5d42fe89-99f8-436c-8401-835532319e14","Type":"ContainerStarted","Data":"bd6d0dc08f638f5a0af576c6533a1799530ed7a79edb144fe33e21a781570240"} Oct 02 10:03:58 crc kubenswrapper[4934]: I1002 10:03:58.041362 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-7c7c8d99ff-g2x7f" Oct 02 10:03:58 crc kubenswrapper[4934]: I1002 10:03:58.069067 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-7c7c8d99ff-g2x7f" podStartSLOduration=1.456449852 podStartE2EDuration="8.069048313s" podCreationTimestamp="2025-10-02 10:03:50 +0000 UTC" firstStartedPulling="2025-10-02 10:03:50.811689147 +0000 UTC m=+902.564330669" lastFinishedPulling="2025-10-02 10:03:57.424287608 +0000 UTC m=+909.176929130" observedRunningTime="2025-10-02 10:03:58.067291745 +0000 UTC m=+909.819933267" watchObservedRunningTime="2025-10-02 10:03:58.069048313 +0000 UTC m=+909.821689845" Oct 02 10:04:00 crc kubenswrapper[4934]: I1002 10:04:00.053104 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-7c7c8d99ff-g2x7f" Oct 02 10:04:08 crc kubenswrapper[4934]: I1002 10:04:08.439359 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:04:08 crc kubenswrapper[4934]: I1002 10:04:08.440120 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.101496 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859cd486d-w2rxq"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.102928 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-w2rxq" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.105154 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-npqcc" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.105512 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-f7f98cb69-lsrzt"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.106630 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-lsrzt" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.107883 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-7f6mf" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.113122 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-77fb7bcf5b-jq9qb"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.114360 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-jq9qb" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.117621 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-h2n9r" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.119197 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-f7f98cb69-lsrzt"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.126281 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-77fb7bcf5b-jq9qb"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.130208 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859cd486d-w2rxq"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.134448 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-8bc4775b5-dh7nc"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.135342 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-dh7nc" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.136787 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-sbfrm" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.177417 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b4fc86755-xf5gv"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.183645 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-xf5gv" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.188034 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-dltbt" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.198835 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-8bc4775b5-dh7nc"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.211144 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-679b4759bb-782sv"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.219697 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-782sv" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.222568 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-4f77p" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.224514 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b4fc86755-xf5gv"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.230371 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-6wzhw"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.231515 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-6wzhw" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.235656 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-679b4759bb-782sv"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.235904 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.236057 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-2qss5" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.243389 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zm95v\" (UniqueName: \"kubernetes.io/projected/5f887690-5ea6-401d-b37c-4623d5911e30-kube-api-access-zm95v\") pod \"designate-operator-controller-manager-77fb7bcf5b-jq9qb\" (UID: \"5f887690-5ea6-401d-b37c-4623d5911e30\") " pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-jq9qb" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.243460 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rlxp\" (UniqueName: \"kubernetes.io/projected/6a0e27a6-27cf-45bf-bb3d-adced5dd2fcd-kube-api-access-7rlxp\") pod \"cinder-operator-controller-manager-859cd486d-w2rxq\" (UID: \"6a0e27a6-27cf-45bf-bb3d-adced5dd2fcd\") " pod="openstack-operators/cinder-operator-controller-manager-859cd486d-w2rxq" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.243489 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2fgx\" (UniqueName: \"kubernetes.io/projected/2b00925c-c02d-4652-ae45-a3eadc31fda5-kube-api-access-c2fgx\") pod \"barbican-operator-controller-manager-f7f98cb69-lsrzt\" (UID: \"2b00925c-c02d-4652-ae45-a3eadc31fda5\") " pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-lsrzt" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.243513 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59lwt\" (UniqueName: \"kubernetes.io/projected/f315b606-7268-48c4-81b8-bb45bc0fb1a9-kube-api-access-59lwt\") pod \"glance-operator-controller-manager-8bc4775b5-dh7nc\" (UID: \"f315b606-7268-48c4-81b8-bb45bc0fb1a9\") " pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-dh7nc" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.252329 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-6wzhw"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.260456 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5f45cd594f-q7drf"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.268843 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-q7drf" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.270286 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-59d7dc95cf-cswb5"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.270422 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-8g8bf" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.271224 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-cswb5" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.276899 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-59d7dc95cf-cswb5"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.277332 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-xndhq" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.281271 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5f45cd594f-q7drf"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.305382 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67bf5bb885-dxd5q"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.306483 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-dxd5q" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.311091 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-b7cf8cb5f-sg4zk"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.314059 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-sg4zk" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.314638 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-qll45" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.320527 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-f4mvb" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.330201 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-54fbbfcd44-2dkpb"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.336554 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-2dkpb" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.344890 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-kl2t9" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.347255 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9v2x\" (UniqueName: \"kubernetes.io/projected/5d8d6294-fa14-4090-a472-862673713bf7-kube-api-access-b9v2x\") pod \"horizon-operator-controller-manager-679b4759bb-782sv\" (UID: \"5d8d6294-fa14-4090-a472-862673713bf7\") " pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-782sv" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.347295 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2fgx\" (UniqueName: \"kubernetes.io/projected/2b00925c-c02d-4652-ae45-a3eadc31fda5-kube-api-access-c2fgx\") pod \"barbican-operator-controller-manager-f7f98cb69-lsrzt\" (UID: \"2b00925c-c02d-4652-ae45-a3eadc31fda5\") " pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-lsrzt" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.347327 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59lwt\" (UniqueName: \"kubernetes.io/projected/f315b606-7268-48c4-81b8-bb45bc0fb1a9-kube-api-access-59lwt\") pod \"glance-operator-controller-manager-8bc4775b5-dh7nc\" (UID: \"f315b606-7268-48c4-81b8-bb45bc0fb1a9\") " pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-dh7nc" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.347356 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ea2637c9-7013-4184-b7e6-3d429181b4aa-cert\") pod \"infra-operator-controller-manager-5c8fdc4d5c-6wzhw\" (UID: \"ea2637c9-7013-4184-b7e6-3d429181b4aa\") " pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-6wzhw" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.347381 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zm95v\" (UniqueName: \"kubernetes.io/projected/5f887690-5ea6-401d-b37c-4623d5911e30-kube-api-access-zm95v\") pod \"designate-operator-controller-manager-77fb7bcf5b-jq9qb\" (UID: \"5f887690-5ea6-401d-b37c-4623d5911e30\") " pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-jq9qb" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.347421 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbfgl\" (UniqueName: \"kubernetes.io/projected/ea2637c9-7013-4184-b7e6-3d429181b4aa-kube-api-access-rbfgl\") pod \"infra-operator-controller-manager-5c8fdc4d5c-6wzhw\" (UID: \"ea2637c9-7013-4184-b7e6-3d429181b4aa\") " pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-6wzhw" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.347453 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldvsz\" (UniqueName: \"kubernetes.io/projected/da0df363-5daf-4162-8f99-93ebd8f9ee24-kube-api-access-ldvsz\") pod \"heat-operator-controller-manager-5b4fc86755-xf5gv\" (UID: \"da0df363-5daf-4162-8f99-93ebd8f9ee24\") " pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-xf5gv" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.347475 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rlxp\" (UniqueName: \"kubernetes.io/projected/6a0e27a6-27cf-45bf-bb3d-adced5dd2fcd-kube-api-access-7rlxp\") pod \"cinder-operator-controller-manager-859cd486d-w2rxq\" (UID: \"6a0e27a6-27cf-45bf-bb3d-adced5dd2fcd\") " pod="openstack-operators/cinder-operator-controller-manager-859cd486d-w2rxq" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.349157 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-54fbbfcd44-2dkpb"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.357643 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-b7cf8cb5f-sg4zk"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.366104 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67bf5bb885-dxd5q"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.378593 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-7wntc"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.379533 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-7wntc" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.381537 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-75f8d67d86-dhf8d"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.382917 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-dhf8d" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.392554 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59lwt\" (UniqueName: \"kubernetes.io/projected/f315b606-7268-48c4-81b8-bb45bc0fb1a9-kube-api-access-59lwt\") pod \"glance-operator-controller-manager-8bc4775b5-dh7nc\" (UID: \"f315b606-7268-48c4-81b8-bb45bc0fb1a9\") " pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-dh7nc" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.395208 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-vwljz" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.395559 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-xht2t" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.399645 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zm95v\" (UniqueName: \"kubernetes.io/projected/5f887690-5ea6-401d-b37c-4623d5911e30-kube-api-access-zm95v\") pod \"designate-operator-controller-manager-77fb7bcf5b-jq9qb\" (UID: \"5f887690-5ea6-401d-b37c-4623d5911e30\") " pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-jq9qb" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.400126 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2fgx\" (UniqueName: \"kubernetes.io/projected/2b00925c-c02d-4652-ae45-a3eadc31fda5-kube-api-access-c2fgx\") pod \"barbican-operator-controller-manager-f7f98cb69-lsrzt\" (UID: \"2b00925c-c02d-4652-ae45-a3eadc31fda5\") " pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-lsrzt" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.402986 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rlxp\" (UniqueName: \"kubernetes.io/projected/6a0e27a6-27cf-45bf-bb3d-adced5dd2fcd-kube-api-access-7rlxp\") pod \"cinder-operator-controller-manager-859cd486d-w2rxq\" (UID: \"6a0e27a6-27cf-45bf-bb3d-adced5dd2fcd\") " pod="openstack-operators/cinder-operator-controller-manager-859cd486d-w2rxq" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.425095 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-w2rxq" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.425504 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-7wntc"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.432932 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-75f8d67d86-dhf8d"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.449031 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ea2637c9-7013-4184-b7e6-3d429181b4aa-cert\") pod \"infra-operator-controller-manager-5c8fdc4d5c-6wzhw\" (UID: \"ea2637c9-7013-4184-b7e6-3d429181b4aa\") " pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-6wzhw" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.449090 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82wp6\" (UniqueName: \"kubernetes.io/projected/d411e947-e43e-4543-b2d7-177639563cc5-kube-api-access-82wp6\") pod \"neutron-operator-controller-manager-54fbbfcd44-2dkpb\" (UID: \"d411e947-e43e-4543-b2d7-177639563cc5\") " pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-2dkpb" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.449131 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glzwm\" (UniqueName: \"kubernetes.io/projected/1e12ae39-0145-42a5-ab78-855441de01fc-kube-api-access-glzwm\") pod \"keystone-operator-controller-manager-59d7dc95cf-cswb5\" (UID: \"1e12ae39-0145-42a5-ab78-855441de01fc\") " pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-cswb5" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.449151 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbfgl\" (UniqueName: \"kubernetes.io/projected/ea2637c9-7013-4184-b7e6-3d429181b4aa-kube-api-access-rbfgl\") pod \"infra-operator-controller-manager-5c8fdc4d5c-6wzhw\" (UID: \"ea2637c9-7013-4184-b7e6-3d429181b4aa\") " pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-6wzhw" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.449171 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldvsz\" (UniqueName: \"kubernetes.io/projected/da0df363-5daf-4162-8f99-93ebd8f9ee24-kube-api-access-ldvsz\") pod \"heat-operator-controller-manager-5b4fc86755-xf5gv\" (UID: \"da0df363-5daf-4162-8f99-93ebd8f9ee24\") " pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-xf5gv" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.449194 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9v2x\" (UniqueName: \"kubernetes.io/projected/5d8d6294-fa14-4090-a472-862673713bf7-kube-api-access-b9v2x\") pod \"horizon-operator-controller-manager-679b4759bb-782sv\" (UID: \"5d8d6294-fa14-4090-a472-862673713bf7\") " pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-782sv" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.449212 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcp4w\" (UniqueName: \"kubernetes.io/projected/06a5a960-2aa6-4b1b-ae51-316873d545f6-kube-api-access-gcp4w\") pod \"mariadb-operator-controller-manager-67bf5bb885-dxd5q\" (UID: \"06a5a960-2aa6-4b1b-ae51-316873d545f6\") " pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-dxd5q" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.449233 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5h4kx\" (UniqueName: \"kubernetes.io/projected/0372f96c-947a-4625-8219-4dda67a14b3f-kube-api-access-5h4kx\") pod \"manila-operator-controller-manager-b7cf8cb5f-sg4zk\" (UID: \"0372f96c-947a-4625-8219-4dda67a14b3f\") " pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-sg4zk" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.449255 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhbcp\" (UniqueName: \"kubernetes.io/projected/d12830ea-c37d-46fe-952b-0d5019faf418-kube-api-access-vhbcp\") pod \"ironic-operator-controller-manager-5f45cd594f-q7drf\" (UID: \"d12830ea-c37d-46fe-952b-0d5019faf418\") " pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-q7drf" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.454810 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-lsrzt" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.466410 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ea2637c9-7013-4184-b7e6-3d429181b4aa-cert\") pod \"infra-operator-controller-manager-5c8fdc4d5c-6wzhw\" (UID: \"ea2637c9-7013-4184-b7e6-3d429181b4aa\") " pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-6wzhw" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.483502 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldvsz\" (UniqueName: \"kubernetes.io/projected/da0df363-5daf-4162-8f99-93ebd8f9ee24-kube-api-access-ldvsz\") pod \"heat-operator-controller-manager-5b4fc86755-xf5gv\" (UID: \"da0df363-5daf-4162-8f99-93ebd8f9ee24\") " pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-xf5gv" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.483566 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.485053 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.486130 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbfgl\" (UniqueName: \"kubernetes.io/projected/ea2637c9-7013-4184-b7e6-3d429181b4aa-kube-api-access-rbfgl\") pod \"infra-operator-controller-manager-5c8fdc4d5c-6wzhw\" (UID: \"ea2637c9-7013-4184-b7e6-3d429181b4aa\") " pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-6wzhw" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.488229 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-jq9qb" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.495879 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-dh7nc" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.506180 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.506431 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-6k5l2" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.512626 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-84c745747f-qj6vc"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.514004 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-qj6vc" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.520093 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9v2x\" (UniqueName: \"kubernetes.io/projected/5d8d6294-fa14-4090-a472-862673713bf7-kube-api-access-b9v2x\") pod \"horizon-operator-controller-manager-679b4759bb-782sv\" (UID: \"5d8d6294-fa14-4090-a472-862673713bf7\") " pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-782sv" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.525376 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-rgpn9" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.525935 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-xf5gv" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.530904 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-598c4c8547-6545m"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.531984 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-6545m" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.537687 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-4fwd8" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.542878 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-84c745747f-qj6vc"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.563121 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-782sv" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.563736 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcp4w\" (UniqueName: \"kubernetes.io/projected/06a5a960-2aa6-4b1b-ae51-316873d545f6-kube-api-access-gcp4w\") pod \"mariadb-operator-controller-manager-67bf5bb885-dxd5q\" (UID: \"06a5a960-2aa6-4b1b-ae51-316873d545f6\") " pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-dxd5q" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.563765 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5h4kx\" (UniqueName: \"kubernetes.io/projected/0372f96c-947a-4625-8219-4dda67a14b3f-kube-api-access-5h4kx\") pod \"manila-operator-controller-manager-b7cf8cb5f-sg4zk\" (UID: \"0372f96c-947a-4625-8219-4dda67a14b3f\") " pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-sg4zk" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.563793 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tq7rs\" (UniqueName: \"kubernetes.io/projected/04d01fc6-fd8c-414d-b8b5-aab49056718d-kube-api-access-tq7rs\") pod \"nova-operator-controller-manager-7fd5b6bbc6-7wntc\" (UID: \"04d01fc6-fd8c-414d-b8b5-aab49056718d\") " pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-7wntc" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.563822 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhbcp\" (UniqueName: \"kubernetes.io/projected/d12830ea-c37d-46fe-952b-0d5019faf418-kube-api-access-vhbcp\") pod \"ironic-operator-controller-manager-5f45cd594f-q7drf\" (UID: \"d12830ea-c37d-46fe-952b-0d5019faf418\") " pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-q7drf" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.563848 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwskt\" (UniqueName: \"kubernetes.io/projected/5df0b2ac-a570-4594-8311-6c682eb07646-kube-api-access-fwskt\") pod \"octavia-operator-controller-manager-75f8d67d86-dhf8d\" (UID: \"5df0b2ac-a570-4594-8311-6c682eb07646\") " pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-dhf8d" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.563894 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82wp6\" (UniqueName: \"kubernetes.io/projected/d411e947-e43e-4543-b2d7-177639563cc5-kube-api-access-82wp6\") pod \"neutron-operator-controller-manager-54fbbfcd44-2dkpb\" (UID: \"d411e947-e43e-4543-b2d7-177639563cc5\") " pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-2dkpb" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.563930 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glzwm\" (UniqueName: \"kubernetes.io/projected/1e12ae39-0145-42a5-ab78-855441de01fc-kube-api-access-glzwm\") pod \"keystone-operator-controller-manager-59d7dc95cf-cswb5\" (UID: \"1e12ae39-0145-42a5-ab78-855441de01fc\") " pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-cswb5" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.565875 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-6wzhw" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.604430 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-689b4f76c9-pk5ms"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.606449 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-pk5ms" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.614600 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5h4kx\" (UniqueName: \"kubernetes.io/projected/0372f96c-947a-4625-8219-4dda67a14b3f-kube-api-access-5h4kx\") pod \"manila-operator-controller-manager-b7cf8cb5f-sg4zk\" (UID: \"0372f96c-947a-4625-8219-4dda67a14b3f\") " pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-sg4zk" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.624121 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcp4w\" (UniqueName: \"kubernetes.io/projected/06a5a960-2aa6-4b1b-ae51-316873d545f6-kube-api-access-gcp4w\") pod \"mariadb-operator-controller-manager-67bf5bb885-dxd5q\" (UID: \"06a5a960-2aa6-4b1b-ae51-316873d545f6\") " pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-dxd5q" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.639689 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhbcp\" (UniqueName: \"kubernetes.io/projected/d12830ea-c37d-46fe-952b-0d5019faf418-kube-api-access-vhbcp\") pod \"ironic-operator-controller-manager-5f45cd594f-q7drf\" (UID: \"d12830ea-c37d-46fe-952b-0d5019faf418\") " pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-q7drf" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.640335 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-dxd5q" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.642095 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-w9wzs" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.646399 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-sg4zk" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.667132 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tq7rs\" (UniqueName: \"kubernetes.io/projected/04d01fc6-fd8c-414d-b8b5-aab49056718d-kube-api-access-tq7rs\") pod \"nova-operator-controller-manager-7fd5b6bbc6-7wntc\" (UID: \"04d01fc6-fd8c-414d-b8b5-aab49056718d\") " pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-7wntc" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.686175 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qb8lp\" (UniqueName: \"kubernetes.io/projected/0fd36c26-564f-4bda-b394-b21c674f5274-kube-api-access-qb8lp\") pod \"placement-operator-controller-manager-598c4c8547-6545m\" (UID: \"0fd36c26-564f-4bda-b394-b21c674f5274\") " pod="openstack-operators/placement-operator-controller-manager-598c4c8547-6545m" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.686655 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwskt\" (UniqueName: \"kubernetes.io/projected/5df0b2ac-a570-4594-8311-6c682eb07646-kube-api-access-fwskt\") pod \"octavia-operator-controller-manager-75f8d67d86-dhf8d\" (UID: \"5df0b2ac-a570-4594-8311-6c682eb07646\") " pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-dhf8d" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.687096 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4f4z\" (UniqueName: \"kubernetes.io/projected/ce5aee9d-cb75-4b25-a795-75a362b717a3-kube-api-access-f4f4z\") pod \"openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg\" (UID: \"ce5aee9d-cb75-4b25-a795-75a362b717a3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.687130 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ce5aee9d-cb75-4b25-a795-75a362b717a3-cert\") pod \"openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg\" (UID: \"ce5aee9d-cb75-4b25-a795-75a362b717a3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.687172 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vx7tk\" (UniqueName: \"kubernetes.io/projected/a93ce93b-0718-4572-997c-3af703524717-kube-api-access-vx7tk\") pod \"ovn-operator-controller-manager-84c745747f-qj6vc\" (UID: \"a93ce93b-0718-4572-997c-3af703524717\") " pod="openstack-operators/ovn-operator-controller-manager-84c745747f-qj6vc" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.688943 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82wp6\" (UniqueName: \"kubernetes.io/projected/d411e947-e43e-4543-b2d7-177639563cc5-kube-api-access-82wp6\") pod \"neutron-operator-controller-manager-54fbbfcd44-2dkpb\" (UID: \"d411e947-e43e-4543-b2d7-177639563cc5\") " pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-2dkpb" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.698211 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.730959 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glzwm\" (UniqueName: \"kubernetes.io/projected/1e12ae39-0145-42a5-ab78-855441de01fc-kube-api-access-glzwm\") pod \"keystone-operator-controller-manager-59d7dc95cf-cswb5\" (UID: \"1e12ae39-0145-42a5-ab78-855441de01fc\") " pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-cswb5" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.732111 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwskt\" (UniqueName: \"kubernetes.io/projected/5df0b2ac-a570-4594-8311-6c682eb07646-kube-api-access-fwskt\") pod \"octavia-operator-controller-manager-75f8d67d86-dhf8d\" (UID: \"5df0b2ac-a570-4594-8311-6c682eb07646\") " pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-dhf8d" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.751221 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-689b4f76c9-pk5ms"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.769088 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tq7rs\" (UniqueName: \"kubernetes.io/projected/04d01fc6-fd8c-414d-b8b5-aab49056718d-kube-api-access-tq7rs\") pod \"nova-operator-controller-manager-7fd5b6bbc6-7wntc\" (UID: \"04d01fc6-fd8c-414d-b8b5-aab49056718d\") " pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-7wntc" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.782620 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-cb66d6b59-zxggl"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.783845 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-zxggl" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.787299 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-vr2vt" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.791522 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-598c4c8547-6545m"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.793631 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qb8lp\" (UniqueName: \"kubernetes.io/projected/0fd36c26-564f-4bda-b394-b21c674f5274-kube-api-access-qb8lp\") pod \"placement-operator-controller-manager-598c4c8547-6545m\" (UID: \"0fd36c26-564f-4bda-b394-b21c674f5274\") " pod="openstack-operators/placement-operator-controller-manager-598c4c8547-6545m" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.793734 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4f4z\" (UniqueName: \"kubernetes.io/projected/ce5aee9d-cb75-4b25-a795-75a362b717a3-kube-api-access-f4f4z\") pod \"openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg\" (UID: \"ce5aee9d-cb75-4b25-a795-75a362b717a3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.793763 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ce5aee9d-cb75-4b25-a795-75a362b717a3-cert\") pod \"openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg\" (UID: \"ce5aee9d-cb75-4b25-a795-75a362b717a3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.793793 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gc6h\" (UniqueName: \"kubernetes.io/projected/50cb1560-d7b8-4149-8fed-931e03814fb5-kube-api-access-7gc6h\") pod \"swift-operator-controller-manager-689b4f76c9-pk5ms\" (UID: \"50cb1560-d7b8-4149-8fed-931e03814fb5\") " pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-pk5ms" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.793824 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vx7tk\" (UniqueName: \"kubernetes.io/projected/a93ce93b-0718-4572-997c-3af703524717-kube-api-access-vx7tk\") pod \"ovn-operator-controller-manager-84c745747f-qj6vc\" (UID: \"a93ce93b-0718-4572-997c-3af703524717\") " pod="openstack-operators/ovn-operator-controller-manager-84c745747f-qj6vc" Oct 02 10:04:28 crc kubenswrapper[4934]: E1002 10:04:28.794358 4934 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 02 10:04:28 crc kubenswrapper[4934]: E1002 10:04:28.794396 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ce5aee9d-cb75-4b25-a795-75a362b717a3-cert podName:ce5aee9d-cb75-4b25-a795-75a362b717a3 nodeName:}" failed. No retries permitted until 2025-10-02 10:04:29.294381626 +0000 UTC m=+941.047023148 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ce5aee9d-cb75-4b25-a795-75a362b717a3-cert") pod "openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg" (UID: "ce5aee9d-cb75-4b25-a795-75a362b717a3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.816939 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-7wntc" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.822988 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vx7tk\" (UniqueName: \"kubernetes.io/projected/a93ce93b-0718-4572-997c-3af703524717-kube-api-access-vx7tk\") pod \"ovn-operator-controller-manager-84c745747f-qj6vc\" (UID: \"a93ce93b-0718-4572-997c-3af703524717\") " pod="openstack-operators/ovn-operator-controller-manager-84c745747f-qj6vc" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.824199 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-cbdf6dc66-xfxx5"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.825461 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-xfxx5" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.828861 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-grr4z" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.829608 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4f4z\" (UniqueName: \"kubernetes.io/projected/ce5aee9d-cb75-4b25-a795-75a362b717a3-kube-api-access-f4f4z\") pod \"openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg\" (UID: \"ce5aee9d-cb75-4b25-a795-75a362b717a3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.832373 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qb8lp\" (UniqueName: \"kubernetes.io/projected/0fd36c26-564f-4bda-b394-b21c674f5274-kube-api-access-qb8lp\") pod \"placement-operator-controller-manager-598c4c8547-6545m\" (UID: \"0fd36c26-564f-4bda-b394-b21c674f5274\") " pod="openstack-operators/placement-operator-controller-manager-598c4c8547-6545m" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.836954 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-cb66d6b59-zxggl"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.865022 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-dhf8d" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.890449 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-q7drf" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.894831 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gc6h\" (UniqueName: \"kubernetes.io/projected/50cb1560-d7b8-4149-8fed-931e03814fb5-kube-api-access-7gc6h\") pod \"swift-operator-controller-manager-689b4f76c9-pk5ms\" (UID: \"50cb1560-d7b8-4149-8fed-931e03814fb5\") " pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-pk5ms" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.894898 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxbnr\" (UniqueName: \"kubernetes.io/projected/dfe7cd03-27c4-4dc5-b875-ca762209e599-kube-api-access-vxbnr\") pod \"telemetry-operator-controller-manager-cb66d6b59-zxggl\" (UID: \"dfe7cd03-27c4-4dc5-b875-ca762209e599\") " pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-zxggl" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.904029 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-cbdf6dc66-xfxx5"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.909888 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-cswb5" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.923663 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-qj6vc" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.933287 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gc6h\" (UniqueName: \"kubernetes.io/projected/50cb1560-d7b8-4149-8fed-931e03814fb5-kube-api-access-7gc6h\") pod \"swift-operator-controller-manager-689b4f76c9-pk5ms\" (UID: \"50cb1560-d7b8-4149-8fed-931e03814fb5\") " pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-pk5ms" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.956115 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-2dkpb" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.971469 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-68d7bc5569-zsmk7"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.972698 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-zsmk7" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.986249 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-6545m" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.986974 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-c8rzq" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.990665 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-68d7bc5569-zsmk7"] Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.999790 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llmwk\" (UniqueName: \"kubernetes.io/projected/0da92003-0ad0-4a25-9c55-16784a8fcc83-kube-api-access-llmwk\") pod \"test-operator-controller-manager-cbdf6dc66-xfxx5\" (UID: \"0da92003-0ad0-4a25-9c55-16784a8fcc83\") " pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-xfxx5" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.999846 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mts6t\" (UniqueName: \"kubernetes.io/projected/f470bad4-e97c-45f2-ac3f-3c3befc41094-kube-api-access-mts6t\") pod \"watcher-operator-controller-manager-68d7bc5569-zsmk7\" (UID: \"f470bad4-e97c-45f2-ac3f-3c3befc41094\") " pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-zsmk7" Oct 02 10:04:28 crc kubenswrapper[4934]: I1002 10:04:28.999952 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxbnr\" (UniqueName: \"kubernetes.io/projected/dfe7cd03-27c4-4dc5-b875-ca762209e599-kube-api-access-vxbnr\") pod \"telemetry-operator-controller-manager-cb66d6b59-zxggl\" (UID: \"dfe7cd03-27c4-4dc5-b875-ca762209e599\") " pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-zxggl" Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.029898 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-794f8d6c6d-ppdh4"] Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.030942 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-794f8d6c6d-ppdh4" Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.034285 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-hmzcn" Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.034439 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.037835 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-794f8d6c6d-ppdh4"] Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.042636 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xbqhj"] Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.043542 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xbqhj" Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.044243 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxbnr\" (UniqueName: \"kubernetes.io/projected/dfe7cd03-27c4-4dc5-b875-ca762209e599-kube-api-access-vxbnr\") pod \"telemetry-operator-controller-manager-cb66d6b59-zxggl\" (UID: \"dfe7cd03-27c4-4dc5-b875-ca762209e599\") " pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-zxggl" Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.048560 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-mdq7v" Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.051106 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xbqhj"] Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.070043 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-pk5ms" Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.100597 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzh6m\" (UniqueName: \"kubernetes.io/projected/d5c07043-ed3a-4dd7-a685-f8dcc7bf578e-kube-api-access-rzh6m\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-xbqhj\" (UID: \"d5c07043-ed3a-4dd7-a685-f8dcc7bf578e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xbqhj" Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.100637 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grzj4\" (UniqueName: \"kubernetes.io/projected/f0e2fb9e-d1cb-42ab-b16b-b64f5e2240e2-kube-api-access-grzj4\") pod \"openstack-operator-controller-manager-794f8d6c6d-ppdh4\" (UID: \"f0e2fb9e-d1cb-42ab-b16b-b64f5e2240e2\") " pod="openstack-operators/openstack-operator-controller-manager-794f8d6c6d-ppdh4" Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.100658 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f0e2fb9e-d1cb-42ab-b16b-b64f5e2240e2-cert\") pod \"openstack-operator-controller-manager-794f8d6c6d-ppdh4\" (UID: \"f0e2fb9e-d1cb-42ab-b16b-b64f5e2240e2\") " pod="openstack-operators/openstack-operator-controller-manager-794f8d6c6d-ppdh4" Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.100712 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llmwk\" (UniqueName: \"kubernetes.io/projected/0da92003-0ad0-4a25-9c55-16784a8fcc83-kube-api-access-llmwk\") pod \"test-operator-controller-manager-cbdf6dc66-xfxx5\" (UID: \"0da92003-0ad0-4a25-9c55-16784a8fcc83\") " pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-xfxx5" Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.100737 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mts6t\" (UniqueName: \"kubernetes.io/projected/f470bad4-e97c-45f2-ac3f-3c3befc41094-kube-api-access-mts6t\") pod \"watcher-operator-controller-manager-68d7bc5569-zsmk7\" (UID: \"f470bad4-e97c-45f2-ac3f-3c3befc41094\") " pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-zsmk7" Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.118642 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llmwk\" (UniqueName: \"kubernetes.io/projected/0da92003-0ad0-4a25-9c55-16784a8fcc83-kube-api-access-llmwk\") pod \"test-operator-controller-manager-cbdf6dc66-xfxx5\" (UID: \"0da92003-0ad0-4a25-9c55-16784a8fcc83\") " pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-xfxx5" Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.118703 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mts6t\" (UniqueName: \"kubernetes.io/projected/f470bad4-e97c-45f2-ac3f-3c3befc41094-kube-api-access-mts6t\") pod \"watcher-operator-controller-manager-68d7bc5569-zsmk7\" (UID: \"f470bad4-e97c-45f2-ac3f-3c3befc41094\") " pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-zsmk7" Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.179855 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-zxggl" Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.201499 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f0e2fb9e-d1cb-42ab-b16b-b64f5e2240e2-cert\") pod \"openstack-operator-controller-manager-794f8d6c6d-ppdh4\" (UID: \"f0e2fb9e-d1cb-42ab-b16b-b64f5e2240e2\") " pod="openstack-operators/openstack-operator-controller-manager-794f8d6c6d-ppdh4" Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.201639 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzh6m\" (UniqueName: \"kubernetes.io/projected/d5c07043-ed3a-4dd7-a685-f8dcc7bf578e-kube-api-access-rzh6m\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-xbqhj\" (UID: \"d5c07043-ed3a-4dd7-a685-f8dcc7bf578e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xbqhj" Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.201671 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grzj4\" (UniqueName: \"kubernetes.io/projected/f0e2fb9e-d1cb-42ab-b16b-b64f5e2240e2-kube-api-access-grzj4\") pod \"openstack-operator-controller-manager-794f8d6c6d-ppdh4\" (UID: \"f0e2fb9e-d1cb-42ab-b16b-b64f5e2240e2\") " pod="openstack-operators/openstack-operator-controller-manager-794f8d6c6d-ppdh4" Oct 02 10:04:29 crc kubenswrapper[4934]: E1002 10:04:29.202036 4934 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 02 10:04:29 crc kubenswrapper[4934]: E1002 10:04:29.202105 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f0e2fb9e-d1cb-42ab-b16b-b64f5e2240e2-cert podName:f0e2fb9e-d1cb-42ab-b16b-b64f5e2240e2 nodeName:}" failed. No retries permitted until 2025-10-02 10:04:29.702084667 +0000 UTC m=+941.454726189 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f0e2fb9e-d1cb-42ab-b16b-b64f5e2240e2-cert") pod "openstack-operator-controller-manager-794f8d6c6d-ppdh4" (UID: "f0e2fb9e-d1cb-42ab-b16b-b64f5e2240e2") : secret "webhook-server-cert" not found Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.219025 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-xfxx5" Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.224285 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grzj4\" (UniqueName: \"kubernetes.io/projected/f0e2fb9e-d1cb-42ab-b16b-b64f5e2240e2-kube-api-access-grzj4\") pod \"openstack-operator-controller-manager-794f8d6c6d-ppdh4\" (UID: \"f0e2fb9e-d1cb-42ab-b16b-b64f5e2240e2\") " pod="openstack-operators/openstack-operator-controller-manager-794f8d6c6d-ppdh4" Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.230007 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzh6m\" (UniqueName: \"kubernetes.io/projected/d5c07043-ed3a-4dd7-a685-f8dcc7bf578e-kube-api-access-rzh6m\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-xbqhj\" (UID: \"d5c07043-ed3a-4dd7-a685-f8dcc7bf578e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xbqhj" Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.302822 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ce5aee9d-cb75-4b25-a795-75a362b717a3-cert\") pod \"openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg\" (UID: \"ce5aee9d-cb75-4b25-a795-75a362b717a3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg" Oct 02 10:04:29 crc kubenswrapper[4934]: E1002 10:04:29.303033 4934 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 02 10:04:29 crc kubenswrapper[4934]: E1002 10:04:29.303091 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ce5aee9d-cb75-4b25-a795-75a362b717a3-cert podName:ce5aee9d-cb75-4b25-a795-75a362b717a3 nodeName:}" failed. No retries permitted until 2025-10-02 10:04:30.303072665 +0000 UTC m=+942.055714187 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ce5aee9d-cb75-4b25-a795-75a362b717a3-cert") pod "openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg" (UID: "ce5aee9d-cb75-4b25-a795-75a362b717a3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.417506 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-zsmk7" Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.441746 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xbqhj" Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.498953 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-f7f98cb69-lsrzt"] Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.515712 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859cd486d-w2rxq"] Oct 02 10:04:29 crc kubenswrapper[4934]: W1002 10:04:29.561355 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b00925c_c02d_4652_ae45_a3eadc31fda5.slice/crio-cb288016c83580089a8498622f6e1f10cd595aeeb9c014072392a462d2849776 WatchSource:0}: Error finding container cb288016c83580089a8498622f6e1f10cd595aeeb9c014072392a462d2849776: Status 404 returned error can't find the container with id cb288016c83580089a8498622f6e1f10cd595aeeb9c014072392a462d2849776 Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.712259 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f0e2fb9e-d1cb-42ab-b16b-b64f5e2240e2-cert\") pod \"openstack-operator-controller-manager-794f8d6c6d-ppdh4\" (UID: \"f0e2fb9e-d1cb-42ab-b16b-b64f5e2240e2\") " pod="openstack-operators/openstack-operator-controller-manager-794f8d6c6d-ppdh4" Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.718036 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f0e2fb9e-d1cb-42ab-b16b-b64f5e2240e2-cert\") pod \"openstack-operator-controller-manager-794f8d6c6d-ppdh4\" (UID: \"f0e2fb9e-d1cb-42ab-b16b-b64f5e2240e2\") " pod="openstack-operators/openstack-operator-controller-manager-794f8d6c6d-ppdh4" Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.730939 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-794f8d6c6d-ppdh4" Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.892958 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-8bc4775b5-dh7nc"] Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.908243 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b4fc86755-xf5gv"] Oct 02 10:04:29 crc kubenswrapper[4934]: W1002 10:04:29.912864 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda0df363_5daf_4162_8f99_93ebd8f9ee24.slice/crio-3e44d0a58d4957e728e7e880faa6229b7b277ae0991567cfb98973329a535516 WatchSource:0}: Error finding container 3e44d0a58d4957e728e7e880faa6229b7b277ae0991567cfb98973329a535516: Status 404 returned error can't find the container with id 3e44d0a58d4957e728e7e880faa6229b7b277ae0991567cfb98973329a535516 Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.914989 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-679b4759bb-782sv"] Oct 02 10:04:29 crc kubenswrapper[4934]: I1002 10:04:29.925817 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-77fb7bcf5b-jq9qb"] Oct 02 10:04:30 crc kubenswrapper[4934]: W1002 10:04:30.171967 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06a5a960_2aa6_4b1b_ae51_316873d545f6.slice/crio-d30cffd8981a318f84e2260b63cd9f940c0100359083a9a3536824c72f98211e WatchSource:0}: Error finding container d30cffd8981a318f84e2260b63cd9f940c0100359083a9a3536824c72f98211e: Status 404 returned error can't find the container with id d30cffd8981a318f84e2260b63cd9f940c0100359083a9a3536824c72f98211e Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.172677 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67bf5bb885-dxd5q"] Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.185280 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-75f8d67d86-dhf8d"] Oct 02 10:04:30 crc kubenswrapper[4934]: W1002 10:04:30.185679 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5df0b2ac_a570_4594_8311_6c682eb07646.slice/crio-2fa254eb0129f46d11f7ea66979fdea96477f828069a9cbd835d6b2fac277e9b WatchSource:0}: Error finding container 2fa254eb0129f46d11f7ea66979fdea96477f828069a9cbd835d6b2fac277e9b: Status 404 returned error can't find the container with id 2fa254eb0129f46d11f7ea66979fdea96477f828069a9cbd835d6b2fac277e9b Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.205623 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-b7cf8cb5f-sg4zk"] Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.219663 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-7wntc"] Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.232643 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-689b4f76c9-pk5ms"] Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.245904 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-6wzhw"] Oct 02 10:04:30 crc kubenswrapper[4934]: W1002 10:04:30.256319 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod04d01fc6_fd8c_414d_b8b5_aab49056718d.slice/crio-00f073cc1e92cf89019f79f119dea4ffc6be25b3c8c986ad12225e06733f2687 WatchSource:0}: Error finding container 00f073cc1e92cf89019f79f119dea4ffc6be25b3c8c986ad12225e06733f2687: Status 404 returned error can't find the container with id 00f073cc1e92cf89019f79f119dea4ffc6be25b3c8c986ad12225e06733f2687 Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.256826 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-54fbbfcd44-2dkpb"] Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.263192 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-84c745747f-qj6vc"] Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.268632 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5f45cd594f-q7drf"] Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.272670 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-598c4c8547-6545m"] Oct 02 10:04:30 crc kubenswrapper[4934]: W1002 10:04:30.273312 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5c07043_ed3a_4dd7_a685_f8dcc7bf578e.slice/crio-3d185720cf2e50918dffe27e36f818a5f17979a4be902e9861f888bfe65e70a6 WatchSource:0}: Error finding container 3d185720cf2e50918dffe27e36f818a5f17979a4be902e9861f888bfe65e70a6: Status 404 returned error can't find the container with id 3d185720cf2e50918dffe27e36f818a5f17979a4be902e9861f888bfe65e70a6 Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.276665 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-59d7dc95cf-cswb5"] Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.277990 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-lsrzt" event={"ID":"2b00925c-c02d-4652-ae45-a3eadc31fda5","Type":"ContainerStarted","Data":"cb288016c83580089a8498622f6e1f10cd595aeeb9c014072392a462d2849776"} Oct 02 10:04:30 crc kubenswrapper[4934]: W1002 10:04:30.278590 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea2637c9_7013_4184_b7e6_3d429181b4aa.slice/crio-ad89c48cb25e5a11be5cd4a700cd8c253f6db358a52eabb45f1f28f66b246d5a WatchSource:0}: Error finding container ad89c48cb25e5a11be5cd4a700cd8c253f6db358a52eabb45f1f28f66b246d5a: Status 404 returned error can't find the container with id ad89c48cb25e5a11be5cd4a700cd8c253f6db358a52eabb45f1f28f66b246d5a Oct 02 10:04:30 crc kubenswrapper[4934]: E1002 10:04:30.283084 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rzh6m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-xbqhj_openstack-operators(d5c07043-ed3a-4dd7-a685-f8dcc7bf578e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 10:04:30 crc kubenswrapper[4934]: E1002 10:04:30.284801 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:ae0bf0aa387864d26d4cbd04d84f6362d53b721348188f096aaa48757419c4f5,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vhbcp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-5f45cd594f-q7drf_openstack-operators(d12830ea-c37d-46fe-952b-0d5019faf418): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 10:04:30 crc kubenswrapper[4934]: E1002 10:04:30.284864 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:3f96f0843934236c261db73dacb50fc12a288890562ee4ebdc9ec22360937cd3,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rbfgl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-5c8fdc4d5c-6wzhw_openstack-operators(ea2637c9-7013-4184-b7e6-3d429181b4aa): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 10:04:30 crc kubenswrapper[4934]: E1002 10:04:30.284926 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xbqhj" podUID="d5c07043-ed3a-4dd7-a685-f8dcc7bf578e" Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.285810 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-sg4zk" event={"ID":"0372f96c-947a-4625-8219-4dda67a14b3f","Type":"ContainerStarted","Data":"2c078eac846636e7293a2a849f7bb52f079f4985058583696f929a2c7d533e36"} Oct 02 10:04:30 crc kubenswrapper[4934]: E1002 10:04:30.285800 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-llmwk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-cbdf6dc66-xfxx5_openstack-operators(0da92003-0ad0-4a25-9c55-16784a8fcc83): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.287806 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-782sv" event={"ID":"5d8d6294-fa14-4090-a472-862673713bf7","Type":"ContainerStarted","Data":"7300ed67d95023da01cd57b9914f9975b3e8da076aac23ef1830189856631ee0"} Oct 02 10:04:30 crc kubenswrapper[4934]: E1002 10:04:30.291115 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:09c2f519ea218f6038b7be039b8e6ac33ee93b217b9be0d2d18a5e7f94faae06,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mts6t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-68d7bc5569-zsmk7_openstack-operators(f470bad4-e97c-45f2-ac3f-3c3befc41094): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.291180 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xbqhj"] Oct 02 10:04:30 crc kubenswrapper[4934]: W1002 10:04:30.297434 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e12ae39_0145_42a5_ab78_855441de01fc.slice/crio-717a34ea88a3e6f103a5a46dcbd841392469d554dfffecde10f5b658be51c088 WatchSource:0}: Error finding container 717a34ea88a3e6f103a5a46dcbd841392469d554dfffecde10f5b658be51c088: Status 404 returned error can't find the container with id 717a34ea88a3e6f103a5a46dcbd841392469d554dfffecde10f5b658be51c088 Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.299698 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-jq9qb" event={"ID":"5f887690-5ea6-401d-b37c-4623d5911e30","Type":"ContainerStarted","Data":"d2a8e20e799812b08121107db0c045c740324647a66f36361ab07d24da6d3bb4"} Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.303282 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-dh7nc" event={"ID":"f315b606-7268-48c4-81b8-bb45bc0fb1a9","Type":"ContainerStarted","Data":"1bbba40960b8cc0e832d8efed294dacfc1454ec73edbb22143296149647048e8"} Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.309209 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-pk5ms" event={"ID":"50cb1560-d7b8-4149-8fed-931e03814fb5","Type":"ContainerStarted","Data":"06ecae0863633c12909931b8a60b36f5bc543821dbad82425c14e9920d679db6"} Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.312166 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-68d7bc5569-zsmk7"] Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.312299 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-7wntc" event={"ID":"04d01fc6-fd8c-414d-b8b5-aab49056718d","Type":"ContainerStarted","Data":"00f073cc1e92cf89019f79f119dea4ffc6be25b3c8c986ad12225e06733f2687"} Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.314415 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-dxd5q" event={"ID":"06a5a960-2aa6-4b1b-ae51-316873d545f6","Type":"ContainerStarted","Data":"d30cffd8981a318f84e2260b63cd9f940c0100359083a9a3536824c72f98211e"} Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.317601 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-xf5gv" event={"ID":"da0df363-5daf-4162-8f99-93ebd8f9ee24","Type":"ContainerStarted","Data":"3e44d0a58d4957e728e7e880faa6229b7b277ae0991567cfb98973329a535516"} Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.318836 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-cb66d6b59-zxggl"] Oct 02 10:04:30 crc kubenswrapper[4934]: E1002 10:04:30.320232 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vxbnr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-cb66d6b59-zxggl_openstack-operators(dfe7cd03-27c4-4dc5-b875-ca762209e599): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 10:04:30 crc kubenswrapper[4934]: E1002 10:04:30.321021 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qb8lp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-598c4c8547-6545m_openstack-operators(0fd36c26-564f-4bda-b394-b21c674f5274): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.321349 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ce5aee9d-cb75-4b25-a795-75a362b717a3-cert\") pod \"openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg\" (UID: \"ce5aee9d-cb75-4b25-a795-75a362b717a3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg" Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.322297 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-w2rxq" event={"ID":"6a0e27a6-27cf-45bf-bb3d-adced5dd2fcd","Type":"ContainerStarted","Data":"2937d22c68458171d2fbdad3f6bcb22c61bbd0592e03f51cbf4c4a447e3bc41f"} Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.323757 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-2dkpb" event={"ID":"d411e947-e43e-4543-b2d7-177639563cc5","Type":"ContainerStarted","Data":"b4f452430809cd0800cebf64a7f04490a0d2f674bec22fc303ab2db4342060f6"} Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.328463 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-cbdf6dc66-xfxx5"] Oct 02 10:04:30 crc kubenswrapper[4934]: E1002 10:04:30.329600 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:23fcec0642cbd40af10bca0c5d4e538662d21eda98d6dfec37c38b4d7a47191a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-glzwm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-59d7dc95cf-cswb5_openstack-operators(1e12ae39-0145-42a5-ab78-855441de01fc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.330617 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-dhf8d" event={"ID":"5df0b2ac-a570-4594-8311-6c682eb07646","Type":"ContainerStarted","Data":"2fa254eb0129f46d11f7ea66979fdea96477f828069a9cbd835d6b2fac277e9b"} Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.330861 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-794f8d6c6d-ppdh4"] Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.331025 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ce5aee9d-cb75-4b25-a795-75a362b717a3-cert\") pod \"openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg\" (UID: \"ce5aee9d-cb75-4b25-a795-75a362b717a3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg" Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.403861 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg" Oct 02 10:04:30 crc kubenswrapper[4934]: E1002 10:04:30.646091 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-xfxx5" podUID="0da92003-0ad0-4a25-9c55-16784a8fcc83" Oct 02 10:04:30 crc kubenswrapper[4934]: E1002 10:04:30.811756 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-6545m" podUID="0fd36c26-564f-4bda-b394-b21c674f5274" Oct 02 10:04:30 crc kubenswrapper[4934]: E1002 10:04:30.831127 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-6wzhw" podUID="ea2637c9-7013-4184-b7e6-3d429181b4aa" Oct 02 10:04:30 crc kubenswrapper[4934]: E1002 10:04:30.831349 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-zsmk7" podUID="f470bad4-e97c-45f2-ac3f-3c3befc41094" Oct 02 10:04:30 crc kubenswrapper[4934]: E1002 10:04:30.831650 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-q7drf" podUID="d12830ea-c37d-46fe-952b-0d5019faf418" Oct 02 10:04:30 crc kubenswrapper[4934]: E1002 10:04:30.832569 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-zxggl" podUID="dfe7cd03-27c4-4dc5-b875-ca762209e599" Oct 02 10:04:30 crc kubenswrapper[4934]: E1002 10:04:30.882654 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-cswb5" podUID="1e12ae39-0145-42a5-ab78-855441de01fc" Oct 02 10:04:30 crc kubenswrapper[4934]: I1002 10:04:30.976415 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg"] Oct 02 10:04:31 crc kubenswrapper[4934]: I1002 10:04:31.338255 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xbqhj" event={"ID":"d5c07043-ed3a-4dd7-a685-f8dcc7bf578e","Type":"ContainerStarted","Data":"3d185720cf2e50918dffe27e36f818a5f17979a4be902e9861f888bfe65e70a6"} Oct 02 10:04:31 crc kubenswrapper[4934]: E1002 10:04:31.343723 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xbqhj" podUID="d5c07043-ed3a-4dd7-a685-f8dcc7bf578e" Oct 02 10:04:31 crc kubenswrapper[4934]: I1002 10:04:31.350522 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-cswb5" event={"ID":"1e12ae39-0145-42a5-ab78-855441de01fc","Type":"ContainerStarted","Data":"4dbacbdf6d7d21e446ed477cf9b0b7d8c084473032b82645fdae284725fc27b5"} Oct 02 10:04:31 crc kubenswrapper[4934]: I1002 10:04:31.350555 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-cswb5" event={"ID":"1e12ae39-0145-42a5-ab78-855441de01fc","Type":"ContainerStarted","Data":"717a34ea88a3e6f103a5a46dcbd841392469d554dfffecde10f5b658be51c088"} Oct 02 10:04:31 crc kubenswrapper[4934]: E1002 10:04:31.354819 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:23fcec0642cbd40af10bca0c5d4e538662d21eda98d6dfec37c38b4d7a47191a\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-cswb5" podUID="1e12ae39-0145-42a5-ab78-855441de01fc" Oct 02 10:04:31 crc kubenswrapper[4934]: I1002 10:04:31.357054 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-6wzhw" event={"ID":"ea2637c9-7013-4184-b7e6-3d429181b4aa","Type":"ContainerStarted","Data":"0b4f10a6ede9ec453b9a2f43cf39191068590ce2bd06445160b500e2e1df5b32"} Oct 02 10:04:31 crc kubenswrapper[4934]: I1002 10:04:31.357111 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-6wzhw" event={"ID":"ea2637c9-7013-4184-b7e6-3d429181b4aa","Type":"ContainerStarted","Data":"ad89c48cb25e5a11be5cd4a700cd8c253f6db358a52eabb45f1f28f66b246d5a"} Oct 02 10:04:31 crc kubenswrapper[4934]: E1002 10:04:31.361504 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:3f96f0843934236c261db73dacb50fc12a288890562ee4ebdc9ec22360937cd3\\\"\"" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-6wzhw" podUID="ea2637c9-7013-4184-b7e6-3d429181b4aa" Oct 02 10:04:31 crc kubenswrapper[4934]: I1002 10:04:31.362173 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-zsmk7" event={"ID":"f470bad4-e97c-45f2-ac3f-3c3befc41094","Type":"ContainerStarted","Data":"c412c2c531d97e6040bd4aa73fd54c7e2b6a9ada7b1250207d8f7f93156acb38"} Oct 02 10:04:31 crc kubenswrapper[4934]: I1002 10:04:31.362208 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-zsmk7" event={"ID":"f470bad4-e97c-45f2-ac3f-3c3befc41094","Type":"ContainerStarted","Data":"908bf2fde23f78cf7d350f6f2877967183ea5ec8333f205352d1ecc5746d1ba0"} Oct 02 10:04:31 crc kubenswrapper[4934]: E1002 10:04:31.364142 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:09c2f519ea218f6038b7be039b8e6ac33ee93b217b9be0d2d18a5e7f94faae06\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-zsmk7" podUID="f470bad4-e97c-45f2-ac3f-3c3befc41094" Oct 02 10:04:31 crc kubenswrapper[4934]: I1002 10:04:31.364404 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-zxggl" event={"ID":"dfe7cd03-27c4-4dc5-b875-ca762209e599","Type":"ContainerStarted","Data":"ab4fade4a703eb536e09285e00511b26b13746aada8436e139e1fd502d2b79c3"} Oct 02 10:04:31 crc kubenswrapper[4934]: I1002 10:04:31.364438 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-zxggl" event={"ID":"dfe7cd03-27c4-4dc5-b875-ca762209e599","Type":"ContainerStarted","Data":"4ec3f0b21cced8164617415fc509d50fc33e569ae04982be4007904c9445bf9b"} Oct 02 10:04:31 crc kubenswrapper[4934]: E1002 10:04:31.367742 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-zxggl" podUID="dfe7cd03-27c4-4dc5-b875-ca762209e599" Oct 02 10:04:31 crc kubenswrapper[4934]: I1002 10:04:31.368388 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-qj6vc" event={"ID":"a93ce93b-0718-4572-997c-3af703524717","Type":"ContainerStarted","Data":"dc3ca8bf6b6140e88452cde8ee6828ac5a082987f24e8e8e556093d7cd6ecb54"} Oct 02 10:04:31 crc kubenswrapper[4934]: I1002 10:04:31.372353 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg" event={"ID":"ce5aee9d-cb75-4b25-a795-75a362b717a3","Type":"ContainerStarted","Data":"921940daf39305cee22d2fe1bd496a7c8e271fc98d6438aedeeaa9c43f1a7451"} Oct 02 10:04:31 crc kubenswrapper[4934]: I1002 10:04:31.376320 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-794f8d6c6d-ppdh4" event={"ID":"f0e2fb9e-d1cb-42ab-b16b-b64f5e2240e2","Type":"ContainerStarted","Data":"019abd0339eabb2805fbf5633d8d7122a4e4ee774f04486e8d7fc0d7fb576eaf"} Oct 02 10:04:31 crc kubenswrapper[4934]: I1002 10:04:31.376373 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-794f8d6c6d-ppdh4" event={"ID":"f0e2fb9e-d1cb-42ab-b16b-b64f5e2240e2","Type":"ContainerStarted","Data":"a019b8cb03fbb175c92b5783f35e616e3e12381ee43615ec303f169b842f0286"} Oct 02 10:04:31 crc kubenswrapper[4934]: I1002 10:04:31.376387 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-794f8d6c6d-ppdh4" event={"ID":"f0e2fb9e-d1cb-42ab-b16b-b64f5e2240e2","Type":"ContainerStarted","Data":"7b71c561da96bf66efb12cd4870109341b979f54a8b77e0bb7360329a5839689"} Oct 02 10:04:31 crc kubenswrapper[4934]: I1002 10:04:31.377163 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-794f8d6c6d-ppdh4" Oct 02 10:04:31 crc kubenswrapper[4934]: I1002 10:04:31.379240 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-q7drf" event={"ID":"d12830ea-c37d-46fe-952b-0d5019faf418","Type":"ContainerStarted","Data":"87d01df50d4470383e4838327e0d04d9bc49abb33577cda436b40e0953363f33"} Oct 02 10:04:31 crc kubenswrapper[4934]: I1002 10:04:31.379265 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-q7drf" event={"ID":"d12830ea-c37d-46fe-952b-0d5019faf418","Type":"ContainerStarted","Data":"2d5518c9e6a7a1dcfccde843a9777353ba48caf109ee4090ad292cbf58a29bef"} Oct 02 10:04:31 crc kubenswrapper[4934]: E1002 10:04:31.382725 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:ae0bf0aa387864d26d4cbd04d84f6362d53b721348188f096aaa48757419c4f5\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-q7drf" podUID="d12830ea-c37d-46fe-952b-0d5019faf418" Oct 02 10:04:31 crc kubenswrapper[4934]: I1002 10:04:31.411387 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-xfxx5" event={"ID":"0da92003-0ad0-4a25-9c55-16784a8fcc83","Type":"ContainerStarted","Data":"27b32cecade28f9b3779081d165288bab833f6f70a8c5fefa26104f81d00862f"} Oct 02 10:04:31 crc kubenswrapper[4934]: I1002 10:04:31.411453 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-xfxx5" event={"ID":"0da92003-0ad0-4a25-9c55-16784a8fcc83","Type":"ContainerStarted","Data":"78634befc144948f52c10d1644627084ab7efb76935c6818d7d76f0bbb113a17"} Oct 02 10:04:31 crc kubenswrapper[4934]: E1002 10:04:31.420407 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3\\\"\"" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-xfxx5" podUID="0da92003-0ad0-4a25-9c55-16784a8fcc83" Oct 02 10:04:31 crc kubenswrapper[4934]: I1002 10:04:31.424817 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-6545m" event={"ID":"0fd36c26-564f-4bda-b394-b21c674f5274","Type":"ContainerStarted","Data":"86393b1dee7ae9d0da4a4509bb80b992375ac1de069ef43f497be1950a7b3f8f"} Oct 02 10:04:31 crc kubenswrapper[4934]: I1002 10:04:31.424862 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-6545m" event={"ID":"0fd36c26-564f-4bda-b394-b21c674f5274","Type":"ContainerStarted","Data":"53de52040871c79ef68d383ece3115922960cc10a9919a04455e0f3515545854"} Oct 02 10:04:31 crc kubenswrapper[4934]: E1002 10:04:31.426514 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2\\\"\"" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-6545m" podUID="0fd36c26-564f-4bda-b394-b21c674f5274" Oct 02 10:04:31 crc kubenswrapper[4934]: I1002 10:04:31.447964 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-794f8d6c6d-ppdh4" podStartSLOduration=3.4479476399999998 podStartE2EDuration="3.44794764s" podCreationTimestamp="2025-10-02 10:04:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:04:31.4461566 +0000 UTC m=+943.198798122" watchObservedRunningTime="2025-10-02 10:04:31.44794764 +0000 UTC m=+943.200589162" Oct 02 10:04:32 crc kubenswrapper[4934]: E1002 10:04:32.442234 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:23fcec0642cbd40af10bca0c5d4e538662d21eda98d6dfec37c38b4d7a47191a\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-cswb5" podUID="1e12ae39-0145-42a5-ab78-855441de01fc" Oct 02 10:04:32 crc kubenswrapper[4934]: E1002 10:04:32.442234 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-zxggl" podUID="dfe7cd03-27c4-4dc5-b875-ca762209e599" Oct 02 10:04:32 crc kubenswrapper[4934]: E1002 10:04:32.442707 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2\\\"\"" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-6545m" podUID="0fd36c26-564f-4bda-b394-b21c674f5274" Oct 02 10:04:32 crc kubenswrapper[4934]: E1002 10:04:32.442834 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:ae0bf0aa387864d26d4cbd04d84f6362d53b721348188f096aaa48757419c4f5\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-q7drf" podUID="d12830ea-c37d-46fe-952b-0d5019faf418" Oct 02 10:04:32 crc kubenswrapper[4934]: E1002 10:04:32.443107 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:3f96f0843934236c261db73dacb50fc12a288890562ee4ebdc9ec22360937cd3\\\"\"" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-6wzhw" podUID="ea2637c9-7013-4184-b7e6-3d429181b4aa" Oct 02 10:04:32 crc kubenswrapper[4934]: E1002 10:04:32.443172 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xbqhj" podUID="d5c07043-ed3a-4dd7-a685-f8dcc7bf578e" Oct 02 10:04:32 crc kubenswrapper[4934]: E1002 10:04:32.444361 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:09c2f519ea218f6038b7be039b8e6ac33ee93b217b9be0d2d18a5e7f94faae06\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-zsmk7" podUID="f470bad4-e97c-45f2-ac3f-3c3befc41094" Oct 02 10:04:32 crc kubenswrapper[4934]: E1002 10:04:32.444416 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3\\\"\"" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-xfxx5" podUID="0da92003-0ad0-4a25-9c55-16784a8fcc83" Oct 02 10:04:38 crc kubenswrapper[4934]: I1002 10:04:38.439595 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:04:38 crc kubenswrapper[4934]: I1002 10:04:38.442064 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:04:38 crc kubenswrapper[4934]: I1002 10:04:38.442446 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 10:04:38 crc kubenswrapper[4934]: I1002 10:04:38.444241 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5567160fec03506e3dac2e60d3a06fa8b4a72b51448a03102f357a2e1f66bb4b"} pod="openshift-machine-config-operator/machine-config-daemon-djh5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 10:04:38 crc kubenswrapper[4934]: I1002 10:04:38.444307 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" containerID="cri-o://5567160fec03506e3dac2e60d3a06fa8b4a72b51448a03102f357a2e1f66bb4b" gracePeriod=600 Oct 02 10:04:39 crc kubenswrapper[4934]: I1002 10:04:39.489535 4934 generic.go:334] "Generic (PLEG): container finished" podID="71db06ef-05b0-4f58-b251-b27117a8500a" containerID="5567160fec03506e3dac2e60d3a06fa8b4a72b51448a03102f357a2e1f66bb4b" exitCode=0 Oct 02 10:04:39 crc kubenswrapper[4934]: I1002 10:04:39.489612 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerDied","Data":"5567160fec03506e3dac2e60d3a06fa8b4a72b51448a03102f357a2e1f66bb4b"} Oct 02 10:04:39 crc kubenswrapper[4934]: I1002 10:04:39.489673 4934 scope.go:117] "RemoveContainer" containerID="3834d18542216db82e6fdc0858cf6c8272b05593949e3231352e2c84300c22de" Oct 02 10:04:39 crc kubenswrapper[4934]: I1002 10:04:39.739353 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-794f8d6c6d-ppdh4" Oct 02 10:04:40 crc kubenswrapper[4934]: I1002 10:04:40.506702 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerStarted","Data":"7b2842dc9e6894e9abab64461b922632ce5e3be7a7134bbbd340310d8f4b1d25"} Oct 02 10:04:40 crc kubenswrapper[4934]: I1002 10:04:40.521221 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-lsrzt" event={"ID":"2b00925c-c02d-4652-ae45-a3eadc31fda5","Type":"ContainerStarted","Data":"079e53e58c80041ec05f6fcf71bfe5ab456f4c6b95b4c19862eb98711bae7b61"} Oct 02 10:04:40 crc kubenswrapper[4934]: I1002 10:04:40.523232 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg" event={"ID":"ce5aee9d-cb75-4b25-a795-75a362b717a3","Type":"ContainerStarted","Data":"088453578ab3971f17a998bd6e3a7834ec9254f5ab43440d435464e51dc95a41"} Oct 02 10:04:40 crc kubenswrapper[4934]: I1002 10:04:40.524684 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-dxd5q" event={"ID":"06a5a960-2aa6-4b1b-ae51-316873d545f6","Type":"ContainerStarted","Data":"55262c2907f63bebcf2e88c566ab59bd99a2671482747c9af059225c9e5930cc"} Oct 02 10:04:40 crc kubenswrapper[4934]: I1002 10:04:40.533564 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-dhf8d" event={"ID":"5df0b2ac-a570-4594-8311-6c682eb07646","Type":"ContainerStarted","Data":"bd63e7e032d412953dedd6e68763a2ce3bcdb6753fc253962293832e295c3c7c"} Oct 02 10:04:40 crc kubenswrapper[4934]: I1002 10:04:40.534733 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-xf5gv" event={"ID":"da0df363-5daf-4162-8f99-93ebd8f9ee24","Type":"ContainerStarted","Data":"8e7f27ee3f37e0a0598883b87c3b9b6e3c9e963a60c37413d871f6af73091a0e"} Oct 02 10:04:40 crc kubenswrapper[4934]: I1002 10:04:40.535913 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-782sv" event={"ID":"5d8d6294-fa14-4090-a472-862673713bf7","Type":"ContainerStarted","Data":"1459461d2b7c15fe97007898670f987cd6fac57ff30848b5b249115439dbcc43"} Oct 02 10:04:40 crc kubenswrapper[4934]: I1002 10:04:40.542206 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-qj6vc" event={"ID":"a93ce93b-0718-4572-997c-3af703524717","Type":"ContainerStarted","Data":"99157137d285d1008a0f9e6262b02f582cc82d89f914b238c9fe53bac963258b"} Oct 02 10:04:40 crc kubenswrapper[4934]: I1002 10:04:40.545146 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-7wntc" event={"ID":"04d01fc6-fd8c-414d-b8b5-aab49056718d","Type":"ContainerStarted","Data":"06377d0b0925bb125b603dc41550fe06ce8baf590312fa5de5035579413958c1"} Oct 02 10:04:40 crc kubenswrapper[4934]: I1002 10:04:40.546262 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-2dkpb" event={"ID":"d411e947-e43e-4543-b2d7-177639563cc5","Type":"ContainerStarted","Data":"6cc688f86e446d8dbda48b8c503fd28b997df2e248b171ba1e98dfac8330aeaa"} Oct 02 10:04:40 crc kubenswrapper[4934]: I1002 10:04:40.547644 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-dh7nc" event={"ID":"f315b606-7268-48c4-81b8-bb45bc0fb1a9","Type":"ContainerStarted","Data":"8d481199da7feb1192f3d8378e0a3c8bef1a607720ab732579db9b392cb5c9bb"} Oct 02 10:04:40 crc kubenswrapper[4934]: I1002 10:04:40.550639 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-pk5ms" event={"ID":"50cb1560-d7b8-4149-8fed-931e03814fb5","Type":"ContainerStarted","Data":"59e1682840fc77f7bd71f5336ebdf3adcdd6b6698f8ba9549a250da35843eb78"} Oct 02 10:04:40 crc kubenswrapper[4934]: I1002 10:04:40.556333 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-w2rxq" event={"ID":"6a0e27a6-27cf-45bf-bb3d-adced5dd2fcd","Type":"ContainerStarted","Data":"127f54660e1a4aaca361a6136fd74cde1714cc7733e14b020e8a049b282a4ac5"} Oct 02 10:04:40 crc kubenswrapper[4934]: I1002 10:04:40.559226 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-sg4zk" event={"ID":"0372f96c-947a-4625-8219-4dda67a14b3f","Type":"ContainerStarted","Data":"a6567f4e3cb055b3baf7c6ac8c9ff10f32410a681c66f11c8657a95a32c6467d"} Oct 02 10:04:40 crc kubenswrapper[4934]: I1002 10:04:40.563417 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-jq9qb" event={"ID":"5f887690-5ea6-401d-b37c-4623d5911e30","Type":"ContainerStarted","Data":"386e13aa3935b1469e34d8bff752e9fd7ac17042536e157b74458307469a391d"} Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.576844 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg" event={"ID":"ce5aee9d-cb75-4b25-a795-75a362b717a3","Type":"ContainerStarted","Data":"7d31a6be2c278642341b7bb265caa5573e1f05879e5604e523152cacf8641a22"} Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.577295 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg" Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.579406 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-7wntc" event={"ID":"04d01fc6-fd8c-414d-b8b5-aab49056718d","Type":"ContainerStarted","Data":"41294bb9a81ecafbd12b9d56489b737f13c0fd71549278a64fda798565a674e7"} Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.579466 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-7wntc" Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.581019 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-lsrzt" event={"ID":"2b00925c-c02d-4652-ae45-a3eadc31fda5","Type":"ContainerStarted","Data":"4f524499b5a1fec5026c3391f3140e3c53dece1adc8495af5fed6fc167a76e57"} Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.581474 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-lsrzt" Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.585124 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-qj6vc" event={"ID":"a93ce93b-0718-4572-997c-3af703524717","Type":"ContainerStarted","Data":"f05abe10816e1d010383bdceea920db9f6b038335885f436f651cd1867419de8"} Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.585354 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-qj6vc" Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.587199 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-jq9qb" event={"ID":"5f887690-5ea6-401d-b37c-4623d5911e30","Type":"ContainerStarted","Data":"d5e99fd4a1696c65d90bec4af5887592e70398e9552b981321c33ec10c257cc4"} Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.587347 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-jq9qb" Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.588914 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-pk5ms" event={"ID":"50cb1560-d7b8-4149-8fed-931e03814fb5","Type":"ContainerStarted","Data":"8fe2081e307891e8af61d7a8a0accd8fe25942fdb2c38821a9f2d3f9446f8b2c"} Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.591082 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-xf5gv" event={"ID":"da0df363-5daf-4162-8f99-93ebd8f9ee24","Type":"ContainerStarted","Data":"ce17184ac7eb05dc2175e4f7211bfb00661cf6b9c9271a3a7555ea7e45e7ba09"} Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.591263 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-xf5gv" Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.595916 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-2dkpb" event={"ID":"d411e947-e43e-4543-b2d7-177639563cc5","Type":"ContainerStarted","Data":"a47caadfe276cf66e15a9514d1160b3312f519745ab319a81dbe66176372a985"} Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.596186 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-2dkpb" Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.597799 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-sg4zk" event={"ID":"0372f96c-947a-4625-8219-4dda67a14b3f","Type":"ContainerStarted","Data":"8429fc6ae98341c0cdddc7a75c6a1d0c9de6973206c5318908276bc66b5acefc"} Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.598349 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-sg4zk" Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.602657 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-dxd5q" event={"ID":"06a5a960-2aa6-4b1b-ae51-316873d545f6","Type":"ContainerStarted","Data":"8894bbdc91ed13b82a663c8b5d535cbdd6df72545b2b846b7d952baa5e87dc0e"} Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.602813 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-dxd5q" Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.604054 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-dhf8d" event={"ID":"5df0b2ac-a570-4594-8311-6c682eb07646","Type":"ContainerStarted","Data":"16f65c61a4eb28b609476cff494e1a736dbdbfc944cc37fb57acd70bfa16f32a"} Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.604189 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-dhf8d" Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.605891 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-dh7nc" event={"ID":"f315b606-7268-48c4-81b8-bb45bc0fb1a9","Type":"ContainerStarted","Data":"47e49c49f49e1d815ff5636076100cc21e9842b66cac1db54c13472ebd53d448"} Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.606035 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-dh7nc" Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.608166 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-w2rxq" event={"ID":"6a0e27a6-27cf-45bf-bb3d-adced5dd2fcd","Type":"ContainerStarted","Data":"a2def024495225e381e3edc6ce5339cc7de6a7445afb5d23c5e8b844c0ab04be"} Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.608246 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-w2rxq" Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.610080 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-782sv" event={"ID":"5d8d6294-fa14-4090-a472-862673713bf7","Type":"ContainerStarted","Data":"a7d97a2661abc872af3750e3a24662c0718d2e551d31631a94442185a9addc4d"} Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.610404 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-782sv" Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.626779 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg" podStartSLOduration=4.869878989 podStartE2EDuration="13.626758357s" podCreationTimestamp="2025-10-02 10:04:28 +0000 UTC" firstStartedPulling="2025-10-02 10:04:30.997341486 +0000 UTC m=+942.749983008" lastFinishedPulling="2025-10-02 10:04:39.754220854 +0000 UTC m=+951.506862376" observedRunningTime="2025-10-02 10:04:41.62232005 +0000 UTC m=+953.374961572" watchObservedRunningTime="2025-10-02 10:04:41.626758357 +0000 UTC m=+953.379399899" Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.671069 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-dh7nc" podStartSLOduration=3.818842402 podStartE2EDuration="13.67104754s" podCreationTimestamp="2025-10-02 10:04:28 +0000 UTC" firstStartedPulling="2025-10-02 10:04:29.900286797 +0000 UTC m=+941.652928319" lastFinishedPulling="2025-10-02 10:04:39.752491935 +0000 UTC m=+951.505133457" observedRunningTime="2025-10-02 10:04:41.652207682 +0000 UTC m=+953.404849214" watchObservedRunningTime="2025-10-02 10:04:41.67104754 +0000 UTC m=+953.423689062" Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.671707 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-pk5ms" podStartSLOduration=4.178869943 podStartE2EDuration="13.671700018s" podCreationTimestamp="2025-10-02 10:04:28 +0000 UTC" firstStartedPulling="2025-10-02 10:04:30.260395141 +0000 UTC m=+942.013036663" lastFinishedPulling="2025-10-02 10:04:39.753225216 +0000 UTC m=+951.505866738" observedRunningTime="2025-10-02 10:04:41.665506362 +0000 UTC m=+953.418147884" watchObservedRunningTime="2025-10-02 10:04:41.671700018 +0000 UTC m=+953.424341540" Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.688241 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-jq9qb" podStartSLOduration=3.883508035 podStartE2EDuration="13.688219549s" podCreationTimestamp="2025-10-02 10:04:28 +0000 UTC" firstStartedPulling="2025-10-02 10:04:29.94988354 +0000 UTC m=+941.702525062" lastFinishedPulling="2025-10-02 10:04:39.754595054 +0000 UTC m=+951.507236576" observedRunningTime="2025-10-02 10:04:41.682881447 +0000 UTC m=+953.435522979" watchObservedRunningTime="2025-10-02 10:04:41.688219549 +0000 UTC m=+953.440861071" Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.710749 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-dhf8d" podStartSLOduration=4.134523771 podStartE2EDuration="13.710732961s" podCreationTimestamp="2025-10-02 10:04:28 +0000 UTC" firstStartedPulling="2025-10-02 10:04:30.195293446 +0000 UTC m=+941.947934968" lastFinishedPulling="2025-10-02 10:04:39.771502636 +0000 UTC m=+951.524144158" observedRunningTime="2025-10-02 10:04:41.707946891 +0000 UTC m=+953.460588423" watchObservedRunningTime="2025-10-02 10:04:41.710732961 +0000 UTC m=+953.463374483" Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.729019 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-dxd5q" podStartSLOduration=4.147671424 podStartE2EDuration="13.728999671s" podCreationTimestamp="2025-10-02 10:04:28 +0000 UTC" firstStartedPulling="2025-10-02 10:04:30.175238394 +0000 UTC m=+941.927879916" lastFinishedPulling="2025-10-02 10:04:39.756566641 +0000 UTC m=+951.509208163" observedRunningTime="2025-10-02 10:04:41.722533177 +0000 UTC m=+953.475174699" watchObservedRunningTime="2025-10-02 10:04:41.728999671 +0000 UTC m=+953.481641193" Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.742318 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-qj6vc" podStartSLOduration=4.325146962 podStartE2EDuration="13.74230351s" podCreationTimestamp="2025-10-02 10:04:28 +0000 UTC" firstStartedPulling="2025-10-02 10:04:30.27756812 +0000 UTC m=+942.030209642" lastFinishedPulling="2025-10-02 10:04:39.694724668 +0000 UTC m=+951.447366190" observedRunningTime="2025-10-02 10:04:41.741120297 +0000 UTC m=+953.493761839" watchObservedRunningTime="2025-10-02 10:04:41.74230351 +0000 UTC m=+953.494945032" Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.772816 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-w2rxq" podStartSLOduration=3.660655602 podStartE2EDuration="13.772793599s" podCreationTimestamp="2025-10-02 10:04:28 +0000 UTC" firstStartedPulling="2025-10-02 10:04:29.562404426 +0000 UTC m=+941.315045948" lastFinishedPulling="2025-10-02 10:04:39.674542383 +0000 UTC m=+951.427183945" observedRunningTime="2025-10-02 10:04:41.757804342 +0000 UTC m=+953.510445864" watchObservedRunningTime="2025-10-02 10:04:41.772793599 +0000 UTC m=+953.525435121" Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.779537 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-lsrzt" podStartSLOduration=3.59039171 podStartE2EDuration="13.779524281s" podCreationTimestamp="2025-10-02 10:04:28 +0000 UTC" firstStartedPulling="2025-10-02 10:04:29.564780924 +0000 UTC m=+941.317422446" lastFinishedPulling="2025-10-02 10:04:39.753913495 +0000 UTC m=+951.506555017" observedRunningTime="2025-10-02 10:04:41.777235825 +0000 UTC m=+953.529877357" watchObservedRunningTime="2025-10-02 10:04:41.779524281 +0000 UTC m=+953.532165803" Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.802417 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-782sv" podStartSLOduration=3.997262526 podStartE2EDuration="13.802392262s" podCreationTimestamp="2025-10-02 10:04:28 +0000 UTC" firstStartedPulling="2025-10-02 10:04:29.950025975 +0000 UTC m=+941.702667507" lastFinishedPulling="2025-10-02 10:04:39.755155721 +0000 UTC m=+951.507797243" observedRunningTime="2025-10-02 10:04:41.797503884 +0000 UTC m=+953.550145406" watchObservedRunningTime="2025-10-02 10:04:41.802392262 +0000 UTC m=+953.555033784" Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.844423 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-xf5gv" podStartSLOduration=4.00787692 podStartE2EDuration="13.84440797s" podCreationTimestamp="2025-10-02 10:04:28 +0000 UTC" firstStartedPulling="2025-10-02 10:04:29.91721634 +0000 UTC m=+941.669857862" lastFinishedPulling="2025-10-02 10:04:39.75374739 +0000 UTC m=+951.506388912" observedRunningTime="2025-10-02 10:04:41.83983777 +0000 UTC m=+953.592479292" watchObservedRunningTime="2025-10-02 10:04:41.84440797 +0000 UTC m=+953.597049492" Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.871675 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-7wntc" podStartSLOduration=4.385163294 podStartE2EDuration="13.871654457s" podCreationTimestamp="2025-10-02 10:04:28 +0000 UTC" firstStartedPulling="2025-10-02 10:04:30.266736213 +0000 UTC m=+942.019377735" lastFinishedPulling="2025-10-02 10:04:39.753227376 +0000 UTC m=+951.505868898" observedRunningTime="2025-10-02 10:04:41.870142964 +0000 UTC m=+953.622784506" watchObservedRunningTime="2025-10-02 10:04:41.871654457 +0000 UTC m=+953.624295989" Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.918188 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-2dkpb" podStartSLOduration=4.454521781 podStartE2EDuration="13.918167783s" podCreationTimestamp="2025-10-02 10:04:28 +0000 UTC" firstStartedPulling="2025-10-02 10:04:30.276818119 +0000 UTC m=+942.029459641" lastFinishedPulling="2025-10-02 10:04:39.740464131 +0000 UTC m=+951.493105643" observedRunningTime="2025-10-02 10:04:41.911455071 +0000 UTC m=+953.664096593" watchObservedRunningTime="2025-10-02 10:04:41.918167783 +0000 UTC m=+953.670809315" Oct 02 10:04:41 crc kubenswrapper[4934]: I1002 10:04:41.919357 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-sg4zk" podStartSLOduration=4.423683211 podStartE2EDuration="13.919348846s" podCreationTimestamp="2025-10-02 10:04:28 +0000 UTC" firstStartedPulling="2025-10-02 10:04:30.260412632 +0000 UTC m=+942.013054154" lastFinishedPulling="2025-10-02 10:04:39.756078267 +0000 UTC m=+951.508719789" observedRunningTime="2025-10-02 10:04:41.893984283 +0000 UTC m=+953.646625815" watchObservedRunningTime="2025-10-02 10:04:41.919348846 +0000 UTC m=+953.671990368" Oct 02 10:04:42 crc kubenswrapper[4934]: I1002 10:04:42.619723 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-pk5ms" Oct 02 10:04:45 crc kubenswrapper[4934]: I1002 10:04:45.647159 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-xfxx5" event={"ID":"0da92003-0ad0-4a25-9c55-16784a8fcc83","Type":"ContainerStarted","Data":"c677ed8a60bfa0a617147b2af7839f528890fd2c14a5b72d02cec7b7de3c94ce"} Oct 02 10:04:45 crc kubenswrapper[4934]: I1002 10:04:45.648308 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-xfxx5" Oct 02 10:04:45 crc kubenswrapper[4934]: I1002 10:04:45.655194 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xbqhj" event={"ID":"d5c07043-ed3a-4dd7-a685-f8dcc7bf578e","Type":"ContainerStarted","Data":"94671ee02e4f08761ed37969356fed6fa7b1a1377f786f2ce693fd619e51b3d2"} Oct 02 10:04:45 crc kubenswrapper[4934]: I1002 10:04:45.666724 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-xfxx5" podStartSLOduration=2.517769397 podStartE2EDuration="17.666687577s" podCreationTimestamp="2025-10-02 10:04:28 +0000 UTC" firstStartedPulling="2025-10-02 10:04:30.285727663 +0000 UTC m=+942.038369185" lastFinishedPulling="2025-10-02 10:04:45.434645843 +0000 UTC m=+957.187287365" observedRunningTime="2025-10-02 10:04:45.661947981 +0000 UTC m=+957.414589503" watchObservedRunningTime="2025-10-02 10:04:45.666687577 +0000 UTC m=+957.419329139" Oct 02 10:04:45 crc kubenswrapper[4934]: I1002 10:04:45.678536 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-xbqhj" podStartSLOduration=2.50804794 podStartE2EDuration="17.678518174s" podCreationTimestamp="2025-10-02 10:04:28 +0000 UTC" firstStartedPulling="2025-10-02 10:04:30.282976135 +0000 UTC m=+942.035617657" lastFinishedPulling="2025-10-02 10:04:45.453446369 +0000 UTC m=+957.206087891" observedRunningTime="2025-10-02 10:04:45.676870916 +0000 UTC m=+957.429512438" watchObservedRunningTime="2025-10-02 10:04:45.678518174 +0000 UTC m=+957.431159696" Oct 02 10:04:47 crc kubenswrapper[4934]: I1002 10:04:47.673776 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-6545m" event={"ID":"0fd36c26-564f-4bda-b394-b21c674f5274","Type":"ContainerStarted","Data":"98c740d8c97694a333463b6d1e272dfcba0ac5bfb702c34ad55473062cdef21f"} Oct 02 10:04:47 crc kubenswrapper[4934]: I1002 10:04:47.674371 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-6545m" Oct 02 10:04:47 crc kubenswrapper[4934]: I1002 10:04:47.693882 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-6545m" podStartSLOduration=3.382795363 podStartE2EDuration="19.693865306s" podCreationTimestamp="2025-10-02 10:04:28 +0000 UTC" firstStartedPulling="2025-10-02 10:04:30.320860705 +0000 UTC m=+942.073502227" lastFinishedPulling="2025-10-02 10:04:46.631930638 +0000 UTC m=+958.384572170" observedRunningTime="2025-10-02 10:04:47.690239953 +0000 UTC m=+959.442881485" watchObservedRunningTime="2025-10-02 10:04:47.693865306 +0000 UTC m=+959.446506828" Oct 02 10:04:48 crc kubenswrapper[4934]: I1002 10:04:48.429039 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859cd486d-w2rxq" Oct 02 10:04:48 crc kubenswrapper[4934]: I1002 10:04:48.468260 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-f7f98cb69-lsrzt" Oct 02 10:04:48 crc kubenswrapper[4934]: I1002 10:04:48.507156 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-77fb7bcf5b-jq9qb" Oct 02 10:04:48 crc kubenswrapper[4934]: I1002 10:04:48.508820 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-8bc4775b5-dh7nc" Oct 02 10:04:48 crc kubenswrapper[4934]: I1002 10:04:48.528419 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5b4fc86755-xf5gv" Oct 02 10:04:48 crc kubenswrapper[4934]: I1002 10:04:48.567471 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-679b4759bb-782sv" Oct 02 10:04:48 crc kubenswrapper[4934]: I1002 10:04:48.642792 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-67bf5bb885-dxd5q" Oct 02 10:04:48 crc kubenswrapper[4934]: I1002 10:04:48.648969 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-b7cf8cb5f-sg4zk" Oct 02 10:04:48 crc kubenswrapper[4934]: I1002 10:04:48.687154 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-zxggl" event={"ID":"dfe7cd03-27c4-4dc5-b875-ca762209e599","Type":"ContainerStarted","Data":"774810bc28bb451200b4b5bf4b344a6c453affb85912f30760192991d831449c"} Oct 02 10:04:48 crc kubenswrapper[4934]: I1002 10:04:48.688215 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-zxggl" Oct 02 10:04:48 crc kubenswrapper[4934]: I1002 10:04:48.709338 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-zxggl" podStartSLOduration=3.120293691 podStartE2EDuration="20.70932014s" podCreationTimestamp="2025-10-02 10:04:28 +0000 UTC" firstStartedPulling="2025-10-02 10:04:30.320121984 +0000 UTC m=+942.072763506" lastFinishedPulling="2025-10-02 10:04:47.909148433 +0000 UTC m=+959.661789955" observedRunningTime="2025-10-02 10:04:48.704258706 +0000 UTC m=+960.456900228" watchObservedRunningTime="2025-10-02 10:04:48.70932014 +0000 UTC m=+960.461961662" Oct 02 10:04:48 crc kubenswrapper[4934]: I1002 10:04:48.823312 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-7fd5b6bbc6-7wntc" Oct 02 10:04:48 crc kubenswrapper[4934]: I1002 10:04:48.867994 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-75f8d67d86-dhf8d" Oct 02 10:04:48 crc kubenswrapper[4934]: I1002 10:04:48.933271 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-84c745747f-qj6vc" Oct 02 10:04:48 crc kubenswrapper[4934]: I1002 10:04:48.963691 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-54fbbfcd44-2dkpb" Oct 02 10:04:49 crc kubenswrapper[4934]: I1002 10:04:49.073097 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-689b4f76c9-pk5ms" Oct 02 10:04:50 crc kubenswrapper[4934]: I1002 10:04:50.409663 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg" Oct 02 10:04:51 crc kubenswrapper[4934]: I1002 10:04:51.706099 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-q7drf" event={"ID":"d12830ea-c37d-46fe-952b-0d5019faf418","Type":"ContainerStarted","Data":"e20ca749fba92e1b0637201a2103d677be4de5fcabda59f45ba3f36e13c43db4"} Oct 02 10:04:51 crc kubenswrapper[4934]: I1002 10:04:51.706884 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-q7drf" Oct 02 10:04:51 crc kubenswrapper[4934]: I1002 10:04:51.708150 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-zsmk7" event={"ID":"f470bad4-e97c-45f2-ac3f-3c3befc41094","Type":"ContainerStarted","Data":"0e641638091a8313fe0b72c576cd5373a45a1ade595c85afc33eedbf957696b8"} Oct 02 10:04:51 crc kubenswrapper[4934]: I1002 10:04:51.708326 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-zsmk7" Oct 02 10:04:51 crc kubenswrapper[4934]: I1002 10:04:51.710845 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-cswb5" event={"ID":"1e12ae39-0145-42a5-ab78-855441de01fc","Type":"ContainerStarted","Data":"47a7b2f9eddb7b813b55310a41a9d252489fdaf4f8c9357c970d92a56c9ce4b3"} Oct 02 10:04:51 crc kubenswrapper[4934]: I1002 10:04:51.711490 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-cswb5" Oct 02 10:04:51 crc kubenswrapper[4934]: I1002 10:04:51.713627 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-6wzhw" event={"ID":"ea2637c9-7013-4184-b7e6-3d429181b4aa","Type":"ContainerStarted","Data":"6032a3c44a39d34270eec1ed9108857174a87a61fa442c1b4cc8cb54ac1dc4d4"} Oct 02 10:04:51 crc kubenswrapper[4934]: I1002 10:04:51.713989 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-6wzhw" Oct 02 10:04:51 crc kubenswrapper[4934]: I1002 10:04:51.725630 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-q7drf" podStartSLOduration=3.041836076 podStartE2EDuration="23.725613844s" podCreationTimestamp="2025-10-02 10:04:28 +0000 UTC" firstStartedPulling="2025-10-02 10:04:30.284645893 +0000 UTC m=+942.037287415" lastFinishedPulling="2025-10-02 10:04:50.968423661 +0000 UTC m=+962.721065183" observedRunningTime="2025-10-02 10:04:51.724707737 +0000 UTC m=+963.477349259" watchObservedRunningTime="2025-10-02 10:04:51.725613844 +0000 UTC m=+963.478255366" Oct 02 10:04:51 crc kubenswrapper[4934]: I1002 10:04:51.743045 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-6wzhw" podStartSLOduration=3.098584893 podStartE2EDuration="23.74302477s" podCreationTimestamp="2025-10-02 10:04:28 +0000 UTC" firstStartedPulling="2025-10-02 10:04:30.284751056 +0000 UTC m=+942.037392578" lastFinishedPulling="2025-10-02 10:04:50.929190933 +0000 UTC m=+962.681832455" observedRunningTime="2025-10-02 10:04:51.741842116 +0000 UTC m=+963.494483638" watchObservedRunningTime="2025-10-02 10:04:51.74302477 +0000 UTC m=+963.495666292" Oct 02 10:04:51 crc kubenswrapper[4934]: I1002 10:04:51.756896 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-zsmk7" podStartSLOduration=3.125212081 podStartE2EDuration="23.756878335s" podCreationTimestamp="2025-10-02 10:04:28 +0000 UTC" firstStartedPulling="2025-10-02 10:04:30.290996933 +0000 UTC m=+942.043638465" lastFinishedPulling="2025-10-02 10:04:50.922663197 +0000 UTC m=+962.675304719" observedRunningTime="2025-10-02 10:04:51.755280489 +0000 UTC m=+963.507922021" watchObservedRunningTime="2025-10-02 10:04:51.756878335 +0000 UTC m=+963.509519857" Oct 02 10:04:51 crc kubenswrapper[4934]: I1002 10:04:51.777542 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-cswb5" podStartSLOduration=3.120157247 podStartE2EDuration="23.777521673s" podCreationTimestamp="2025-10-02 10:04:28 +0000 UTC" firstStartedPulling="2025-10-02 10:04:30.32946808 +0000 UTC m=+942.082109602" lastFinishedPulling="2025-10-02 10:04:50.986832506 +0000 UTC m=+962.739474028" observedRunningTime="2025-10-02 10:04:51.773485228 +0000 UTC m=+963.526126750" watchObservedRunningTime="2025-10-02 10:04:51.777521673 +0000 UTC m=+963.530163195" Oct 02 10:04:58 crc kubenswrapper[4934]: I1002 10:04:58.572890 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-5c8fdc4d5c-6wzhw" Oct 02 10:04:58 crc kubenswrapper[4934]: I1002 10:04:58.896431 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5f45cd594f-q7drf" Oct 02 10:04:58 crc kubenswrapper[4934]: I1002 10:04:58.922623 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-59d7dc95cf-cswb5" Oct 02 10:04:58 crc kubenswrapper[4934]: I1002 10:04:58.989827 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-598c4c8547-6545m" Oct 02 10:04:59 crc kubenswrapper[4934]: I1002 10:04:59.184218 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-cb66d6b59-zxggl" Oct 02 10:04:59 crc kubenswrapper[4934]: I1002 10:04:59.224191 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-cbdf6dc66-xfxx5" Oct 02 10:04:59 crc kubenswrapper[4934]: I1002 10:04:59.421412 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-68d7bc5569-zsmk7" Oct 02 10:05:14 crc kubenswrapper[4934]: I1002 10:05:14.771842 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7968455899-qq7k2"] Oct 02 10:05:14 crc kubenswrapper[4934]: I1002 10:05:14.773339 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7968455899-qq7k2" Oct 02 10:05:14 crc kubenswrapper[4934]: I1002 10:05:14.775519 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 02 10:05:14 crc kubenswrapper[4934]: I1002 10:05:14.775565 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-spdjc" Oct 02 10:05:14 crc kubenswrapper[4934]: I1002 10:05:14.775981 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 02 10:05:14 crc kubenswrapper[4934]: I1002 10:05:14.776032 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 02 10:05:14 crc kubenswrapper[4934]: I1002 10:05:14.792299 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7968455899-qq7k2"] Oct 02 10:05:14 crc kubenswrapper[4934]: I1002 10:05:14.823105 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjgwg\" (UniqueName: \"kubernetes.io/projected/dc548c73-564f-4830-a6f1-faa1912d39e6-kube-api-access-bjgwg\") pod \"dnsmasq-dns-7968455899-qq7k2\" (UID: \"dc548c73-564f-4830-a6f1-faa1912d39e6\") " pod="openstack/dnsmasq-dns-7968455899-qq7k2" Oct 02 10:05:14 crc kubenswrapper[4934]: I1002 10:05:14.823593 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc548c73-564f-4830-a6f1-faa1912d39e6-config\") pod \"dnsmasq-dns-7968455899-qq7k2\" (UID: \"dc548c73-564f-4830-a6f1-faa1912d39e6\") " pod="openstack/dnsmasq-dns-7968455899-qq7k2" Oct 02 10:05:14 crc kubenswrapper[4934]: I1002 10:05:14.837950 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-574db999b5-fw6n9"] Oct 02 10:05:14 crc kubenswrapper[4934]: I1002 10:05:14.839122 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-574db999b5-fw6n9" Oct 02 10:05:14 crc kubenswrapper[4934]: I1002 10:05:14.844600 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 02 10:05:14 crc kubenswrapper[4934]: I1002 10:05:14.846975 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-574db999b5-fw6n9"] Oct 02 10:05:14 crc kubenswrapper[4934]: I1002 10:05:14.925876 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b952bbc6-20b0-45b3-90ce-04e7ef92a79c-dns-svc\") pod \"dnsmasq-dns-574db999b5-fw6n9\" (UID: \"b952bbc6-20b0-45b3-90ce-04e7ef92a79c\") " pod="openstack/dnsmasq-dns-574db999b5-fw6n9" Oct 02 10:05:14 crc kubenswrapper[4934]: I1002 10:05:14.925972 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b952bbc6-20b0-45b3-90ce-04e7ef92a79c-config\") pod \"dnsmasq-dns-574db999b5-fw6n9\" (UID: \"b952bbc6-20b0-45b3-90ce-04e7ef92a79c\") " pod="openstack/dnsmasq-dns-574db999b5-fw6n9" Oct 02 10:05:14 crc kubenswrapper[4934]: I1002 10:05:14.926040 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjgwg\" (UniqueName: \"kubernetes.io/projected/dc548c73-564f-4830-a6f1-faa1912d39e6-kube-api-access-bjgwg\") pod \"dnsmasq-dns-7968455899-qq7k2\" (UID: \"dc548c73-564f-4830-a6f1-faa1912d39e6\") " pod="openstack/dnsmasq-dns-7968455899-qq7k2" Oct 02 10:05:14 crc kubenswrapper[4934]: I1002 10:05:14.926095 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s85s2\" (UniqueName: \"kubernetes.io/projected/b952bbc6-20b0-45b3-90ce-04e7ef92a79c-kube-api-access-s85s2\") pod \"dnsmasq-dns-574db999b5-fw6n9\" (UID: \"b952bbc6-20b0-45b3-90ce-04e7ef92a79c\") " pod="openstack/dnsmasq-dns-574db999b5-fw6n9" Oct 02 10:05:14 crc kubenswrapper[4934]: I1002 10:05:14.926145 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc548c73-564f-4830-a6f1-faa1912d39e6-config\") pod \"dnsmasq-dns-7968455899-qq7k2\" (UID: \"dc548c73-564f-4830-a6f1-faa1912d39e6\") " pod="openstack/dnsmasq-dns-7968455899-qq7k2" Oct 02 10:05:14 crc kubenswrapper[4934]: I1002 10:05:14.927304 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc548c73-564f-4830-a6f1-faa1912d39e6-config\") pod \"dnsmasq-dns-7968455899-qq7k2\" (UID: \"dc548c73-564f-4830-a6f1-faa1912d39e6\") " pod="openstack/dnsmasq-dns-7968455899-qq7k2" Oct 02 10:05:14 crc kubenswrapper[4934]: I1002 10:05:14.946345 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjgwg\" (UniqueName: \"kubernetes.io/projected/dc548c73-564f-4830-a6f1-faa1912d39e6-kube-api-access-bjgwg\") pod \"dnsmasq-dns-7968455899-qq7k2\" (UID: \"dc548c73-564f-4830-a6f1-faa1912d39e6\") " pod="openstack/dnsmasq-dns-7968455899-qq7k2" Oct 02 10:05:15 crc kubenswrapper[4934]: I1002 10:05:15.028423 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s85s2\" (UniqueName: \"kubernetes.io/projected/b952bbc6-20b0-45b3-90ce-04e7ef92a79c-kube-api-access-s85s2\") pod \"dnsmasq-dns-574db999b5-fw6n9\" (UID: \"b952bbc6-20b0-45b3-90ce-04e7ef92a79c\") " pod="openstack/dnsmasq-dns-574db999b5-fw6n9" Oct 02 10:05:15 crc kubenswrapper[4934]: I1002 10:05:15.028552 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b952bbc6-20b0-45b3-90ce-04e7ef92a79c-dns-svc\") pod \"dnsmasq-dns-574db999b5-fw6n9\" (UID: \"b952bbc6-20b0-45b3-90ce-04e7ef92a79c\") " pod="openstack/dnsmasq-dns-574db999b5-fw6n9" Oct 02 10:05:15 crc kubenswrapper[4934]: I1002 10:05:15.028710 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b952bbc6-20b0-45b3-90ce-04e7ef92a79c-config\") pod \"dnsmasq-dns-574db999b5-fw6n9\" (UID: \"b952bbc6-20b0-45b3-90ce-04e7ef92a79c\") " pod="openstack/dnsmasq-dns-574db999b5-fw6n9" Oct 02 10:05:15 crc kubenswrapper[4934]: I1002 10:05:15.030273 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b952bbc6-20b0-45b3-90ce-04e7ef92a79c-config\") pod \"dnsmasq-dns-574db999b5-fw6n9\" (UID: \"b952bbc6-20b0-45b3-90ce-04e7ef92a79c\") " pod="openstack/dnsmasq-dns-574db999b5-fw6n9" Oct 02 10:05:15 crc kubenswrapper[4934]: I1002 10:05:15.030813 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b952bbc6-20b0-45b3-90ce-04e7ef92a79c-dns-svc\") pod \"dnsmasq-dns-574db999b5-fw6n9\" (UID: \"b952bbc6-20b0-45b3-90ce-04e7ef92a79c\") " pod="openstack/dnsmasq-dns-574db999b5-fw6n9" Oct 02 10:05:15 crc kubenswrapper[4934]: I1002 10:05:15.048486 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s85s2\" (UniqueName: \"kubernetes.io/projected/b952bbc6-20b0-45b3-90ce-04e7ef92a79c-kube-api-access-s85s2\") pod \"dnsmasq-dns-574db999b5-fw6n9\" (UID: \"b952bbc6-20b0-45b3-90ce-04e7ef92a79c\") " pod="openstack/dnsmasq-dns-574db999b5-fw6n9" Oct 02 10:05:15 crc kubenswrapper[4934]: I1002 10:05:15.090117 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7968455899-qq7k2" Oct 02 10:05:15 crc kubenswrapper[4934]: I1002 10:05:15.153815 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-574db999b5-fw6n9" Oct 02 10:05:15 crc kubenswrapper[4934]: I1002 10:05:15.589506 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7968455899-qq7k2"] Oct 02 10:05:15 crc kubenswrapper[4934]: W1002 10:05:15.599460 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc548c73_564f_4830_a6f1_faa1912d39e6.slice/crio-11fbbbe18c5a52b50153581f88ee9726b0be8f47698ce3e8ca4747b60f8c9e9f WatchSource:0}: Error finding container 11fbbbe18c5a52b50153581f88ee9726b0be8f47698ce3e8ca4747b60f8c9e9f: Status 404 returned error can't find the container with id 11fbbbe18c5a52b50153581f88ee9726b0be8f47698ce3e8ca4747b60f8c9e9f Oct 02 10:05:15 crc kubenswrapper[4934]: I1002 10:05:15.602829 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 10:05:15 crc kubenswrapper[4934]: I1002 10:05:15.647392 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-574db999b5-fw6n9"] Oct 02 10:05:15 crc kubenswrapper[4934]: W1002 10:05:15.652797 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb952bbc6_20b0_45b3_90ce_04e7ef92a79c.slice/crio-1869ec405c80fde61d2320a26a2c0e850449f7d7319c87c6d30b367b10f16120 WatchSource:0}: Error finding container 1869ec405c80fde61d2320a26a2c0e850449f7d7319c87c6d30b367b10f16120: Status 404 returned error can't find the container with id 1869ec405c80fde61d2320a26a2c0e850449f7d7319c87c6d30b367b10f16120 Oct 02 10:05:15 crc kubenswrapper[4934]: I1002 10:05:15.895911 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-574db999b5-fw6n9" event={"ID":"b952bbc6-20b0-45b3-90ce-04e7ef92a79c","Type":"ContainerStarted","Data":"1869ec405c80fde61d2320a26a2c0e850449f7d7319c87c6d30b367b10f16120"} Oct 02 10:05:15 crc kubenswrapper[4934]: I1002 10:05:15.897696 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7968455899-qq7k2" event={"ID":"dc548c73-564f-4830-a6f1-faa1912d39e6","Type":"ContainerStarted","Data":"11fbbbe18c5a52b50153581f88ee9726b0be8f47698ce3e8ca4747b60f8c9e9f"} Oct 02 10:05:16 crc kubenswrapper[4934]: I1002 10:05:16.250711 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-574db999b5-fw6n9"] Oct 02 10:05:16 crc kubenswrapper[4934]: I1002 10:05:16.273849 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7878bd5965-n75cl"] Oct 02 10:05:16 crc kubenswrapper[4934]: I1002 10:05:16.275213 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7878bd5965-n75cl" Oct 02 10:05:16 crc kubenswrapper[4934]: I1002 10:05:16.289296 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7878bd5965-n75cl"] Oct 02 10:05:16 crc kubenswrapper[4934]: I1002 10:05:16.348425 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1ed141e-7d82-47c6-8d17-c1a1e1d601a7-config\") pod \"dnsmasq-dns-7878bd5965-n75cl\" (UID: \"a1ed141e-7d82-47c6-8d17-c1a1e1d601a7\") " pod="openstack/dnsmasq-dns-7878bd5965-n75cl" Oct 02 10:05:16 crc kubenswrapper[4934]: I1002 10:05:16.348524 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4hrl\" (UniqueName: \"kubernetes.io/projected/a1ed141e-7d82-47c6-8d17-c1a1e1d601a7-kube-api-access-b4hrl\") pod \"dnsmasq-dns-7878bd5965-n75cl\" (UID: \"a1ed141e-7d82-47c6-8d17-c1a1e1d601a7\") " pod="openstack/dnsmasq-dns-7878bd5965-n75cl" Oct 02 10:05:16 crc kubenswrapper[4934]: I1002 10:05:16.348572 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1ed141e-7d82-47c6-8d17-c1a1e1d601a7-dns-svc\") pod \"dnsmasq-dns-7878bd5965-n75cl\" (UID: \"a1ed141e-7d82-47c6-8d17-c1a1e1d601a7\") " pod="openstack/dnsmasq-dns-7878bd5965-n75cl" Oct 02 10:05:16 crc kubenswrapper[4934]: I1002 10:05:16.449974 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4hrl\" (UniqueName: \"kubernetes.io/projected/a1ed141e-7d82-47c6-8d17-c1a1e1d601a7-kube-api-access-b4hrl\") pod \"dnsmasq-dns-7878bd5965-n75cl\" (UID: \"a1ed141e-7d82-47c6-8d17-c1a1e1d601a7\") " pod="openstack/dnsmasq-dns-7878bd5965-n75cl" Oct 02 10:05:16 crc kubenswrapper[4934]: I1002 10:05:16.450041 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1ed141e-7d82-47c6-8d17-c1a1e1d601a7-dns-svc\") pod \"dnsmasq-dns-7878bd5965-n75cl\" (UID: \"a1ed141e-7d82-47c6-8d17-c1a1e1d601a7\") " pod="openstack/dnsmasq-dns-7878bd5965-n75cl" Oct 02 10:05:16 crc kubenswrapper[4934]: I1002 10:05:16.450069 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1ed141e-7d82-47c6-8d17-c1a1e1d601a7-config\") pod \"dnsmasq-dns-7878bd5965-n75cl\" (UID: \"a1ed141e-7d82-47c6-8d17-c1a1e1d601a7\") " pod="openstack/dnsmasq-dns-7878bd5965-n75cl" Oct 02 10:05:16 crc kubenswrapper[4934]: I1002 10:05:16.451195 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1ed141e-7d82-47c6-8d17-c1a1e1d601a7-config\") pod \"dnsmasq-dns-7878bd5965-n75cl\" (UID: \"a1ed141e-7d82-47c6-8d17-c1a1e1d601a7\") " pod="openstack/dnsmasq-dns-7878bd5965-n75cl" Oct 02 10:05:16 crc kubenswrapper[4934]: I1002 10:05:16.452108 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1ed141e-7d82-47c6-8d17-c1a1e1d601a7-dns-svc\") pod \"dnsmasq-dns-7878bd5965-n75cl\" (UID: \"a1ed141e-7d82-47c6-8d17-c1a1e1d601a7\") " pod="openstack/dnsmasq-dns-7878bd5965-n75cl" Oct 02 10:05:16 crc kubenswrapper[4934]: I1002 10:05:16.491118 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4hrl\" (UniqueName: \"kubernetes.io/projected/a1ed141e-7d82-47c6-8d17-c1a1e1d601a7-kube-api-access-b4hrl\") pod \"dnsmasq-dns-7878bd5965-n75cl\" (UID: \"a1ed141e-7d82-47c6-8d17-c1a1e1d601a7\") " pod="openstack/dnsmasq-dns-7878bd5965-n75cl" Oct 02 10:05:16 crc kubenswrapper[4934]: I1002 10:05:16.624148 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7878bd5965-n75cl" Oct 02 10:05:16 crc kubenswrapper[4934]: I1002 10:05:16.838317 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7968455899-qq7k2"] Oct 02 10:05:16 crc kubenswrapper[4934]: I1002 10:05:16.847826 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b5fff88f5-9wtdr"] Oct 02 10:05:16 crc kubenswrapper[4934]: I1002 10:05:16.849104 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b5fff88f5-9wtdr" Oct 02 10:05:16 crc kubenswrapper[4934]: I1002 10:05:16.872239 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b5fff88f5-9wtdr"] Oct 02 10:05:16 crc kubenswrapper[4934]: I1002 10:05:16.963263 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knlzs\" (UniqueName: \"kubernetes.io/projected/d8d5c788-c201-4daa-9b3c-3ea12cf696ff-kube-api-access-knlzs\") pod \"dnsmasq-dns-6b5fff88f5-9wtdr\" (UID: \"d8d5c788-c201-4daa-9b3c-3ea12cf696ff\") " pod="openstack/dnsmasq-dns-6b5fff88f5-9wtdr" Oct 02 10:05:16 crc kubenswrapper[4934]: I1002 10:05:16.963320 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8d5c788-c201-4daa-9b3c-3ea12cf696ff-config\") pod \"dnsmasq-dns-6b5fff88f5-9wtdr\" (UID: \"d8d5c788-c201-4daa-9b3c-3ea12cf696ff\") " pod="openstack/dnsmasq-dns-6b5fff88f5-9wtdr" Oct 02 10:05:16 crc kubenswrapper[4934]: I1002 10:05:16.963342 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8d5c788-c201-4daa-9b3c-3ea12cf696ff-dns-svc\") pod \"dnsmasq-dns-6b5fff88f5-9wtdr\" (UID: \"d8d5c788-c201-4daa-9b3c-3ea12cf696ff\") " pod="openstack/dnsmasq-dns-6b5fff88f5-9wtdr" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.064981 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knlzs\" (UniqueName: \"kubernetes.io/projected/d8d5c788-c201-4daa-9b3c-3ea12cf696ff-kube-api-access-knlzs\") pod \"dnsmasq-dns-6b5fff88f5-9wtdr\" (UID: \"d8d5c788-c201-4daa-9b3c-3ea12cf696ff\") " pod="openstack/dnsmasq-dns-6b5fff88f5-9wtdr" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.065366 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8d5c788-c201-4daa-9b3c-3ea12cf696ff-config\") pod \"dnsmasq-dns-6b5fff88f5-9wtdr\" (UID: \"d8d5c788-c201-4daa-9b3c-3ea12cf696ff\") " pod="openstack/dnsmasq-dns-6b5fff88f5-9wtdr" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.065415 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8d5c788-c201-4daa-9b3c-3ea12cf696ff-dns-svc\") pod \"dnsmasq-dns-6b5fff88f5-9wtdr\" (UID: \"d8d5c788-c201-4daa-9b3c-3ea12cf696ff\") " pod="openstack/dnsmasq-dns-6b5fff88f5-9wtdr" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.067261 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8d5c788-c201-4daa-9b3c-3ea12cf696ff-config\") pod \"dnsmasq-dns-6b5fff88f5-9wtdr\" (UID: \"d8d5c788-c201-4daa-9b3c-3ea12cf696ff\") " pod="openstack/dnsmasq-dns-6b5fff88f5-9wtdr" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.067274 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8d5c788-c201-4daa-9b3c-3ea12cf696ff-dns-svc\") pod \"dnsmasq-dns-6b5fff88f5-9wtdr\" (UID: \"d8d5c788-c201-4daa-9b3c-3ea12cf696ff\") " pod="openstack/dnsmasq-dns-6b5fff88f5-9wtdr" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.082139 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knlzs\" (UniqueName: \"kubernetes.io/projected/d8d5c788-c201-4daa-9b3c-3ea12cf696ff-kube-api-access-knlzs\") pod \"dnsmasq-dns-6b5fff88f5-9wtdr\" (UID: \"d8d5c788-c201-4daa-9b3c-3ea12cf696ff\") " pod="openstack/dnsmasq-dns-6b5fff88f5-9wtdr" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.161616 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7878bd5965-n75cl"] Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.189647 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b5fff88f5-9wtdr" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.444453 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.451097 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.453867 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-769kp" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.457988 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.458134 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.458245 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.458339 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.458484 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.458624 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.470436 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d2b6e083-a523-48e2-8157-088036174e12-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.470470 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d2b6e083-a523-48e2-8157-088036174e12-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.470496 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d2b6e083-a523-48e2-8157-088036174e12-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.470514 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.470537 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d2b6e083-a523-48e2-8157-088036174e12-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.470630 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mqz7\" (UniqueName: \"kubernetes.io/projected/d2b6e083-a523-48e2-8157-088036174e12-kube-api-access-6mqz7\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.470659 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d2b6e083-a523-48e2-8157-088036174e12-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.470679 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d2b6e083-a523-48e2-8157-088036174e12-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.470700 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d2b6e083-a523-48e2-8157-088036174e12-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.470723 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d2b6e083-a523-48e2-8157-088036174e12-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.470738 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d2b6e083-a523-48e2-8157-088036174e12-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.473230 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.571892 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mqz7\" (UniqueName: \"kubernetes.io/projected/d2b6e083-a523-48e2-8157-088036174e12-kube-api-access-6mqz7\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.571954 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d2b6e083-a523-48e2-8157-088036174e12-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.571972 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d2b6e083-a523-48e2-8157-088036174e12-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.571992 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d2b6e083-a523-48e2-8157-088036174e12-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.572018 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d2b6e083-a523-48e2-8157-088036174e12-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.572035 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d2b6e083-a523-48e2-8157-088036174e12-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.572680 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d2b6e083-a523-48e2-8157-088036174e12-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.572723 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d2b6e083-a523-48e2-8157-088036174e12-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.572751 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d2b6e083-a523-48e2-8157-088036174e12-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.572771 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.572806 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d2b6e083-a523-48e2-8157-088036174e12-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.573714 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.591747 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b5fff88f5-9wtdr"] Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.626055 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d2b6e083-a523-48e2-8157-088036174e12-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.626945 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d2b6e083-a523-48e2-8157-088036174e12-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.631057 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d2b6e083-a523-48e2-8157-088036174e12-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: W1002 10:05:17.634462 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8d5c788_c201_4daa_9b3c_3ea12cf696ff.slice/crio-09b782665b0760fdbfbdfd37dca5dc87592b65a07d1bd5b9d4141ac729c271a0 WatchSource:0}: Error finding container 09b782665b0760fdbfbdfd37dca5dc87592b65a07d1bd5b9d4141ac729c271a0: Status 404 returned error can't find the container with id 09b782665b0760fdbfbdfd37dca5dc87592b65a07d1bd5b9d4141ac729c271a0 Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.636324 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d2b6e083-a523-48e2-8157-088036174e12-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.636852 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mqz7\" (UniqueName: \"kubernetes.io/projected/d2b6e083-a523-48e2-8157-088036174e12-kube-api-access-6mqz7\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.642439 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.642517 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d2b6e083-a523-48e2-8157-088036174e12-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.642714 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d2b6e083-a523-48e2-8157-088036174e12-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.642738 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d2b6e083-a523-48e2-8157-088036174e12-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.644294 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d2b6e083-a523-48e2-8157-088036174e12-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.646916 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d2b6e083-a523-48e2-8157-088036174e12-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.836602 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.935187 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7878bd5965-n75cl" event={"ID":"a1ed141e-7d82-47c6-8d17-c1a1e1d601a7","Type":"ContainerStarted","Data":"a58c1061224f039e736f8aaef018231338d07eaa3725105d31a0338383520abc"} Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.937293 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b5fff88f5-9wtdr" event={"ID":"d8d5c788-c201-4daa-9b3c-3ea12cf696ff","Type":"ContainerStarted","Data":"09b782665b0760fdbfbdfd37dca5dc87592b65a07d1bd5b9d4141ac729c271a0"} Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.988892 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.991804 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.994971 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.995141 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.995243 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-sslbg" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.996010 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 02 10:05:17 crc kubenswrapper[4934]: I1002 10:05:17.996293 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:17.999977 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:17.999990 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.008275 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.185368 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sk6pk\" (UniqueName: \"kubernetes.io/projected/97560469-c459-48f4-901e-54eaded32537-kube-api-access-sk6pk\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.185698 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/97560469-c459-48f4-901e-54eaded32537-pod-info\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.185732 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/97560469-c459-48f4-901e-54eaded32537-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.185748 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/97560469-c459-48f4-901e-54eaded32537-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.185772 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/97560469-c459-48f4-901e-54eaded32537-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.185791 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/97560469-c459-48f4-901e-54eaded32537-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.186156 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/97560469-c459-48f4-901e-54eaded32537-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.186204 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/97560469-c459-48f4-901e-54eaded32537-config-data\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.186249 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/97560469-c459-48f4-901e-54eaded32537-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.186281 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.186309 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/97560469-c459-48f4-901e-54eaded32537-server-conf\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.254642 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 10:05:18 crc kubenswrapper[4934]: W1002 10:05:18.261367 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd2b6e083_a523_48e2_8157_088036174e12.slice/crio-035be76bc25d2b585700ddf5bdb5f97eee1b3039ef484bb89d5941763c64ad15 WatchSource:0}: Error finding container 035be76bc25d2b585700ddf5bdb5f97eee1b3039ef484bb89d5941763c64ad15: Status 404 returned error can't find the container with id 035be76bc25d2b585700ddf5bdb5f97eee1b3039ef484bb89d5941763c64ad15 Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.290089 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sk6pk\" (UniqueName: \"kubernetes.io/projected/97560469-c459-48f4-901e-54eaded32537-kube-api-access-sk6pk\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.290170 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/97560469-c459-48f4-901e-54eaded32537-pod-info\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.290210 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/97560469-c459-48f4-901e-54eaded32537-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.290260 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/97560469-c459-48f4-901e-54eaded32537-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.290297 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/97560469-c459-48f4-901e-54eaded32537-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.290326 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/97560469-c459-48f4-901e-54eaded32537-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.290363 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/97560469-c459-48f4-901e-54eaded32537-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.290392 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/97560469-c459-48f4-901e-54eaded32537-config-data\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.290437 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/97560469-c459-48f4-901e-54eaded32537-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.290465 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.290549 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/97560469-c459-48f4-901e-54eaded32537-server-conf\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.291083 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.291655 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/97560469-c459-48f4-901e-54eaded32537-config-data\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.291795 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/97560469-c459-48f4-901e-54eaded32537-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.292462 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/97560469-c459-48f4-901e-54eaded32537-server-conf\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.292765 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/97560469-c459-48f4-901e-54eaded32537-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.293077 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/97560469-c459-48f4-901e-54eaded32537-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.295022 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/97560469-c459-48f4-901e-54eaded32537-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.295357 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/97560469-c459-48f4-901e-54eaded32537-pod-info\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.300841 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/97560469-c459-48f4-901e-54eaded32537-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.308123 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sk6pk\" (UniqueName: \"kubernetes.io/projected/97560469-c459-48f4-901e-54eaded32537-kube-api-access-sk6pk\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.310312 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/97560469-c459-48f4-901e-54eaded32537-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.332426 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.621336 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 10:05:18 crc kubenswrapper[4934]: I1002 10:05:18.956824 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d2b6e083-a523-48e2-8157-088036174e12","Type":"ContainerStarted","Data":"035be76bc25d2b585700ddf5bdb5f97eee1b3039ef484bb89d5941763c64ad15"} Oct 02 10:05:19 crc kubenswrapper[4934]: I1002 10:05:19.064358 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 10:05:19 crc kubenswrapper[4934]: W1002 10:05:19.070571 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97560469_c459_48f4_901e_54eaded32537.slice/crio-d45171f7a1d489fdaf8f0fc59fb7c0b5a8d0522ee54658cd489af97652b84e4b WatchSource:0}: Error finding container d45171f7a1d489fdaf8f0fc59fb7c0b5a8d0522ee54658cd489af97652b84e4b: Status 404 returned error can't find the container with id d45171f7a1d489fdaf8f0fc59fb7c0b5a8d0522ee54658cd489af97652b84e4b Oct 02 10:05:19 crc kubenswrapper[4934]: I1002 10:05:19.979046 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"97560469-c459-48f4-901e-54eaded32537","Type":"ContainerStarted","Data":"d45171f7a1d489fdaf8f0fc59fb7c0b5a8d0522ee54658cd489af97652b84e4b"} Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.231059 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.232280 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.237221 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.238398 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-sbtdf" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.239309 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.239387 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.239550 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.245672 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.260815 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.428511 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-config-data-default\") pod \"openstack-galera-0\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.428555 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76gt4\" (UniqueName: \"kubernetes.io/projected/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-kube-api-access-76gt4\") pod \"openstack-galera-0\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.428640 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.428677 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-secrets\") pod \"openstack-galera-0\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.428690 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-operator-scripts\") pod \"openstack-galera-0\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.428707 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.428871 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.428946 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-kolla-config\") pod \"openstack-galera-0\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.429016 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-config-data-generated\") pod \"openstack-galera-0\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.530612 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-secrets\") pod \"openstack-galera-0\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.530657 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-operator-scripts\") pod \"openstack-galera-0\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.530684 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.530752 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.530796 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-kolla-config\") pod \"openstack-galera-0\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.530850 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-config-data-generated\") pod \"openstack-galera-0\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.530891 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-config-data-default\") pod \"openstack-galera-0\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.530912 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76gt4\" (UniqueName: \"kubernetes.io/projected/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-kube-api-access-76gt4\") pod \"openstack-galera-0\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.530930 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.532054 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-operator-scripts\") pod \"openstack-galera-0\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.532230 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-kolla-config\") pod \"openstack-galera-0\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.532516 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.533189 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-config-data-generated\") pod \"openstack-galera-0\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.533825 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-config-data-default\") pod \"openstack-galera-0\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.537843 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.538095 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-secrets\") pod \"openstack-galera-0\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.538282 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.548809 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76gt4\" (UniqueName: \"kubernetes.io/projected/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-kube-api-access-76gt4\") pod \"openstack-galera-0\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.561277 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.565441 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.636524 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.646152 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.649136 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-xbpbz" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.649346 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.651750 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.651928 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.696165 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.836127 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/125b40d7-516c-432f-ae51-703f2b18068d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.836227 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.836259 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/125b40d7-516c-432f-ae51-703f2b18068d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.836279 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/125b40d7-516c-432f-ae51-703f2b18068d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.836319 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgqm8\" (UniqueName: \"kubernetes.io/projected/125b40d7-516c-432f-ae51-703f2b18068d-kube-api-access-bgqm8\") pod \"openstack-cell1-galera-0\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.836351 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/125b40d7-516c-432f-ae51-703f2b18068d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.836371 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/125b40d7-516c-432f-ae51-703f2b18068d-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.836399 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/125b40d7-516c-432f-ae51-703f2b18068d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.836418 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/125b40d7-516c-432f-ae51-703f2b18068d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.937779 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/125b40d7-516c-432f-ae51-703f2b18068d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.937840 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/125b40d7-516c-432f-ae51-703f2b18068d-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.937880 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/125b40d7-516c-432f-ae51-703f2b18068d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.937905 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/125b40d7-516c-432f-ae51-703f2b18068d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.937942 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/125b40d7-516c-432f-ae51-703f2b18068d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.937992 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.938028 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/125b40d7-516c-432f-ae51-703f2b18068d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.938049 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/125b40d7-516c-432f-ae51-703f2b18068d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.938086 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgqm8\" (UniqueName: \"kubernetes.io/projected/125b40d7-516c-432f-ae51-703f2b18068d-kube-api-access-bgqm8\") pod \"openstack-cell1-galera-0\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.939003 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/125b40d7-516c-432f-ae51-703f2b18068d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.939141 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.939818 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/125b40d7-516c-432f-ae51-703f2b18068d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.942415 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/125b40d7-516c-432f-ae51-703f2b18068d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.945782 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/125b40d7-516c-432f-ae51-703f2b18068d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.948038 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/125b40d7-516c-432f-ae51-703f2b18068d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.949238 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/125b40d7-516c-432f-ae51-703f2b18068d-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.950875 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/125b40d7-516c-432f-ae51-703f2b18068d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.965531 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.982847 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgqm8\" (UniqueName: \"kubernetes.io/projected/125b40d7-516c-432f-ae51-703f2b18068d-kube-api-access-bgqm8\") pod \"openstack-cell1-galera-0\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:20 crc kubenswrapper[4934]: I1002 10:05:20.997387 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:21 crc kubenswrapper[4934]: I1002 10:05:21.111609 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 02 10:05:21 crc kubenswrapper[4934]: I1002 10:05:21.112503 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 02 10:05:21 crc kubenswrapper[4934]: I1002 10:05:21.115984 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 02 10:05:21 crc kubenswrapper[4934]: I1002 10:05:21.116150 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-qtvnw" Oct 02 10:05:21 crc kubenswrapper[4934]: I1002 10:05:21.116279 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 02 10:05:21 crc kubenswrapper[4934]: I1002 10:05:21.120092 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 02 10:05:21 crc kubenswrapper[4934]: I1002 10:05:21.246926 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jvtp\" (UniqueName: \"kubernetes.io/projected/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-kube-api-access-6jvtp\") pod \"memcached-0\" (UID: \"f43ed4d4-3637-4ef6-a914-3a751cba8b4e\") " pod="openstack/memcached-0" Oct 02 10:05:21 crc kubenswrapper[4934]: I1002 10:05:21.246986 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-config-data\") pod \"memcached-0\" (UID: \"f43ed4d4-3637-4ef6-a914-3a751cba8b4e\") " pod="openstack/memcached-0" Oct 02 10:05:21 crc kubenswrapper[4934]: I1002 10:05:21.247063 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"f43ed4d4-3637-4ef6-a914-3a751cba8b4e\") " pod="openstack/memcached-0" Oct 02 10:05:21 crc kubenswrapper[4934]: I1002 10:05:21.247165 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"f43ed4d4-3637-4ef6-a914-3a751cba8b4e\") " pod="openstack/memcached-0" Oct 02 10:05:21 crc kubenswrapper[4934]: I1002 10:05:21.247214 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-kolla-config\") pod \"memcached-0\" (UID: \"f43ed4d4-3637-4ef6-a914-3a751cba8b4e\") " pod="openstack/memcached-0" Oct 02 10:05:21 crc kubenswrapper[4934]: I1002 10:05:21.349343 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jvtp\" (UniqueName: \"kubernetes.io/projected/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-kube-api-access-6jvtp\") pod \"memcached-0\" (UID: \"f43ed4d4-3637-4ef6-a914-3a751cba8b4e\") " pod="openstack/memcached-0" Oct 02 10:05:21 crc kubenswrapper[4934]: I1002 10:05:21.349413 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-config-data\") pod \"memcached-0\" (UID: \"f43ed4d4-3637-4ef6-a914-3a751cba8b4e\") " pod="openstack/memcached-0" Oct 02 10:05:21 crc kubenswrapper[4934]: I1002 10:05:21.349891 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"f43ed4d4-3637-4ef6-a914-3a751cba8b4e\") " pod="openstack/memcached-0" Oct 02 10:05:21 crc kubenswrapper[4934]: I1002 10:05:21.349941 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"f43ed4d4-3637-4ef6-a914-3a751cba8b4e\") " pod="openstack/memcached-0" Oct 02 10:05:21 crc kubenswrapper[4934]: I1002 10:05:21.349967 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-kolla-config\") pod \"memcached-0\" (UID: \"f43ed4d4-3637-4ef6-a914-3a751cba8b4e\") " pod="openstack/memcached-0" Oct 02 10:05:21 crc kubenswrapper[4934]: I1002 10:05:21.350298 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-config-data\") pod \"memcached-0\" (UID: \"f43ed4d4-3637-4ef6-a914-3a751cba8b4e\") " pod="openstack/memcached-0" Oct 02 10:05:21 crc kubenswrapper[4934]: I1002 10:05:21.350750 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-kolla-config\") pod \"memcached-0\" (UID: \"f43ed4d4-3637-4ef6-a914-3a751cba8b4e\") " pod="openstack/memcached-0" Oct 02 10:05:21 crc kubenswrapper[4934]: I1002 10:05:21.355352 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"f43ed4d4-3637-4ef6-a914-3a751cba8b4e\") " pod="openstack/memcached-0" Oct 02 10:05:21 crc kubenswrapper[4934]: I1002 10:05:21.364123 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"f43ed4d4-3637-4ef6-a914-3a751cba8b4e\") " pod="openstack/memcached-0" Oct 02 10:05:21 crc kubenswrapper[4934]: I1002 10:05:21.387119 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jvtp\" (UniqueName: \"kubernetes.io/projected/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-kube-api-access-6jvtp\") pod \"memcached-0\" (UID: \"f43ed4d4-3637-4ef6-a914-3a751cba8b4e\") " pod="openstack/memcached-0" Oct 02 10:05:21 crc kubenswrapper[4934]: I1002 10:05:21.428928 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 02 10:05:23 crc kubenswrapper[4934]: I1002 10:05:23.018365 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 10:05:23 crc kubenswrapper[4934]: I1002 10:05:23.019570 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 10:05:23 crc kubenswrapper[4934]: I1002 10:05:23.025004 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-5th5x" Oct 02 10:05:23 crc kubenswrapper[4934]: I1002 10:05:23.036360 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 10:05:23 crc kubenswrapper[4934]: I1002 10:05:23.180544 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9brt\" (UniqueName: \"kubernetes.io/projected/0d5014eb-503f-45ea-bb78-bc5978dbaa69-kube-api-access-s9brt\") pod \"kube-state-metrics-0\" (UID: \"0d5014eb-503f-45ea-bb78-bc5978dbaa69\") " pod="openstack/kube-state-metrics-0" Oct 02 10:05:23 crc kubenswrapper[4934]: I1002 10:05:23.281804 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9brt\" (UniqueName: \"kubernetes.io/projected/0d5014eb-503f-45ea-bb78-bc5978dbaa69-kube-api-access-s9brt\") pod \"kube-state-metrics-0\" (UID: \"0d5014eb-503f-45ea-bb78-bc5978dbaa69\") " pod="openstack/kube-state-metrics-0" Oct 02 10:05:23 crc kubenswrapper[4934]: I1002 10:05:23.297427 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9brt\" (UniqueName: \"kubernetes.io/projected/0d5014eb-503f-45ea-bb78-bc5978dbaa69-kube-api-access-s9brt\") pod \"kube-state-metrics-0\" (UID: \"0d5014eb-503f-45ea-bb78-bc5978dbaa69\") " pod="openstack/kube-state-metrics-0" Oct 02 10:05:23 crc kubenswrapper[4934]: I1002 10:05:23.339595 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.154258 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-gwpkm"] Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.155524 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gwpkm" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.157624 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.157663 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-cnxpf" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.157845 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.167888 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gwpkm"] Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.179212 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-65wtc"] Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.181085 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-65wtc" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.214670 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-65wtc"] Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.330246 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7214291c-3c8e-4173-a6fd-ef16170258e2-combined-ca-bundle\") pod \"ovn-controller-gwpkm\" (UID: \"7214291c-3c8e-4173-a6fd-ef16170258e2\") " pod="openstack/ovn-controller-gwpkm" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.330287 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qj9h7\" (UniqueName: \"kubernetes.io/projected/7214291c-3c8e-4173-a6fd-ef16170258e2-kube-api-access-qj9h7\") pod \"ovn-controller-gwpkm\" (UID: \"7214291c-3c8e-4173-a6fd-ef16170258e2\") " pod="openstack/ovn-controller-gwpkm" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.330324 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-etc-ovs\") pod \"ovn-controller-ovs-65wtc\" (UID: \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\") " pod="openstack/ovn-controller-ovs-65wtc" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.330349 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-scripts\") pod \"ovn-controller-ovs-65wtc\" (UID: \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\") " pod="openstack/ovn-controller-ovs-65wtc" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.330369 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-var-lib\") pod \"ovn-controller-ovs-65wtc\" (UID: \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\") " pod="openstack/ovn-controller-ovs-65wtc" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.330396 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7214291c-3c8e-4173-a6fd-ef16170258e2-var-run-ovn\") pod \"ovn-controller-gwpkm\" (UID: \"7214291c-3c8e-4173-a6fd-ef16170258e2\") " pod="openstack/ovn-controller-gwpkm" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.330425 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/7214291c-3c8e-4173-a6fd-ef16170258e2-ovn-controller-tls-certs\") pod \"ovn-controller-gwpkm\" (UID: \"7214291c-3c8e-4173-a6fd-ef16170258e2\") " pod="openstack/ovn-controller-gwpkm" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.330443 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-var-log\") pod \"ovn-controller-ovs-65wtc\" (UID: \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\") " pod="openstack/ovn-controller-ovs-65wtc" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.330463 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7214291c-3c8e-4173-a6fd-ef16170258e2-scripts\") pod \"ovn-controller-gwpkm\" (UID: \"7214291c-3c8e-4173-a6fd-ef16170258e2\") " pod="openstack/ovn-controller-gwpkm" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.330480 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8x84\" (UniqueName: \"kubernetes.io/projected/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-kube-api-access-x8x84\") pod \"ovn-controller-ovs-65wtc\" (UID: \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\") " pod="openstack/ovn-controller-ovs-65wtc" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.330494 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7214291c-3c8e-4173-a6fd-ef16170258e2-var-log-ovn\") pod \"ovn-controller-gwpkm\" (UID: \"7214291c-3c8e-4173-a6fd-ef16170258e2\") " pod="openstack/ovn-controller-gwpkm" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.330508 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-var-run\") pod \"ovn-controller-ovs-65wtc\" (UID: \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\") " pod="openstack/ovn-controller-ovs-65wtc" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.330535 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7214291c-3c8e-4173-a6fd-ef16170258e2-var-run\") pod \"ovn-controller-gwpkm\" (UID: \"7214291c-3c8e-4173-a6fd-ef16170258e2\") " pod="openstack/ovn-controller-gwpkm" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.431546 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7214291c-3c8e-4173-a6fd-ef16170258e2-combined-ca-bundle\") pod \"ovn-controller-gwpkm\" (UID: \"7214291c-3c8e-4173-a6fd-ef16170258e2\") " pod="openstack/ovn-controller-gwpkm" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.431602 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qj9h7\" (UniqueName: \"kubernetes.io/projected/7214291c-3c8e-4173-a6fd-ef16170258e2-kube-api-access-qj9h7\") pod \"ovn-controller-gwpkm\" (UID: \"7214291c-3c8e-4173-a6fd-ef16170258e2\") " pod="openstack/ovn-controller-gwpkm" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.431636 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-etc-ovs\") pod \"ovn-controller-ovs-65wtc\" (UID: \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\") " pod="openstack/ovn-controller-ovs-65wtc" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.431664 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-scripts\") pod \"ovn-controller-ovs-65wtc\" (UID: \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\") " pod="openstack/ovn-controller-ovs-65wtc" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.431687 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-var-lib\") pod \"ovn-controller-ovs-65wtc\" (UID: \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\") " pod="openstack/ovn-controller-ovs-65wtc" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.431712 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7214291c-3c8e-4173-a6fd-ef16170258e2-var-run-ovn\") pod \"ovn-controller-gwpkm\" (UID: \"7214291c-3c8e-4173-a6fd-ef16170258e2\") " pod="openstack/ovn-controller-gwpkm" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.431741 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/7214291c-3c8e-4173-a6fd-ef16170258e2-ovn-controller-tls-certs\") pod \"ovn-controller-gwpkm\" (UID: \"7214291c-3c8e-4173-a6fd-ef16170258e2\") " pod="openstack/ovn-controller-gwpkm" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.431759 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-var-log\") pod \"ovn-controller-ovs-65wtc\" (UID: \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\") " pod="openstack/ovn-controller-ovs-65wtc" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.431779 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7214291c-3c8e-4173-a6fd-ef16170258e2-scripts\") pod \"ovn-controller-gwpkm\" (UID: \"7214291c-3c8e-4173-a6fd-ef16170258e2\") " pod="openstack/ovn-controller-gwpkm" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.431794 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8x84\" (UniqueName: \"kubernetes.io/projected/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-kube-api-access-x8x84\") pod \"ovn-controller-ovs-65wtc\" (UID: \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\") " pod="openstack/ovn-controller-ovs-65wtc" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.431808 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7214291c-3c8e-4173-a6fd-ef16170258e2-var-log-ovn\") pod \"ovn-controller-gwpkm\" (UID: \"7214291c-3c8e-4173-a6fd-ef16170258e2\") " pod="openstack/ovn-controller-gwpkm" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.431824 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-var-run\") pod \"ovn-controller-ovs-65wtc\" (UID: \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\") " pod="openstack/ovn-controller-ovs-65wtc" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.431851 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7214291c-3c8e-4173-a6fd-ef16170258e2-var-run\") pod \"ovn-controller-gwpkm\" (UID: \"7214291c-3c8e-4173-a6fd-ef16170258e2\") " pod="openstack/ovn-controller-gwpkm" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.432492 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-var-log\") pod \"ovn-controller-ovs-65wtc\" (UID: \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\") " pod="openstack/ovn-controller-ovs-65wtc" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.432527 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-etc-ovs\") pod \"ovn-controller-ovs-65wtc\" (UID: \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\") " pod="openstack/ovn-controller-ovs-65wtc" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.432668 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7214291c-3c8e-4173-a6fd-ef16170258e2-var-log-ovn\") pod \"ovn-controller-gwpkm\" (UID: \"7214291c-3c8e-4173-a6fd-ef16170258e2\") " pod="openstack/ovn-controller-gwpkm" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.432676 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-var-run\") pod \"ovn-controller-ovs-65wtc\" (UID: \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\") " pod="openstack/ovn-controller-ovs-65wtc" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.432711 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7214291c-3c8e-4173-a6fd-ef16170258e2-var-run\") pod \"ovn-controller-gwpkm\" (UID: \"7214291c-3c8e-4173-a6fd-ef16170258e2\") " pod="openstack/ovn-controller-gwpkm" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.432832 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7214291c-3c8e-4173-a6fd-ef16170258e2-var-run-ovn\") pod \"ovn-controller-gwpkm\" (UID: \"7214291c-3c8e-4173-a6fd-ef16170258e2\") " pod="openstack/ovn-controller-gwpkm" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.432889 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-var-lib\") pod \"ovn-controller-ovs-65wtc\" (UID: \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\") " pod="openstack/ovn-controller-ovs-65wtc" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.434552 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-scripts\") pod \"ovn-controller-ovs-65wtc\" (UID: \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\") " pod="openstack/ovn-controller-ovs-65wtc" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.435213 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7214291c-3c8e-4173-a6fd-ef16170258e2-scripts\") pod \"ovn-controller-gwpkm\" (UID: \"7214291c-3c8e-4173-a6fd-ef16170258e2\") " pod="openstack/ovn-controller-gwpkm" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.438528 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/7214291c-3c8e-4173-a6fd-ef16170258e2-ovn-controller-tls-certs\") pod \"ovn-controller-gwpkm\" (UID: \"7214291c-3c8e-4173-a6fd-ef16170258e2\") " pod="openstack/ovn-controller-gwpkm" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.438619 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7214291c-3c8e-4173-a6fd-ef16170258e2-combined-ca-bundle\") pod \"ovn-controller-gwpkm\" (UID: \"7214291c-3c8e-4173-a6fd-ef16170258e2\") " pod="openstack/ovn-controller-gwpkm" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.448372 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8x84\" (UniqueName: \"kubernetes.io/projected/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-kube-api-access-x8x84\") pod \"ovn-controller-ovs-65wtc\" (UID: \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\") " pod="openstack/ovn-controller-ovs-65wtc" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.454109 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qj9h7\" (UniqueName: \"kubernetes.io/projected/7214291c-3c8e-4173-a6fd-ef16170258e2-kube-api-access-qj9h7\") pod \"ovn-controller-gwpkm\" (UID: \"7214291c-3c8e-4173-a6fd-ef16170258e2\") " pod="openstack/ovn-controller-gwpkm" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.472503 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gwpkm" Oct 02 10:05:26 crc kubenswrapper[4934]: I1002 10:05:26.496397 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-65wtc" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.541645 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.543902 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.547300 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.549059 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.549207 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.549337 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-dsxhq" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.549490 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.550793 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.651762 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7tqx\" (UniqueName: \"kubernetes.io/projected/9d816754-2d4c-40e6-acee-218947ab9e19-kube-api-access-s7tqx\") pod \"ovsdbserver-nb-0\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.651823 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9d816754-2d4c-40e6-acee-218947ab9e19-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.651846 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d816754-2d4c-40e6-acee-218947ab9e19-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.651877 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d816754-2d4c-40e6-acee-218947ab9e19-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.651896 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d816754-2d4c-40e6-acee-218947ab9e19-config\") pod \"ovsdbserver-nb-0\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.651914 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d816754-2d4c-40e6-acee-218947ab9e19-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.651935 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d816754-2d4c-40e6-acee-218947ab9e19-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.652117 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.754662 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d816754-2d4c-40e6-acee-218947ab9e19-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.754778 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d816754-2d4c-40e6-acee-218947ab9e19-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.754835 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.754994 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7tqx\" (UniqueName: \"kubernetes.io/projected/9d816754-2d4c-40e6-acee-218947ab9e19-kube-api-access-s7tqx\") pod \"ovsdbserver-nb-0\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.755096 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9d816754-2d4c-40e6-acee-218947ab9e19-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.755125 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d816754-2d4c-40e6-acee-218947ab9e19-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.755177 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d816754-2d4c-40e6-acee-218947ab9e19-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.755200 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d816754-2d4c-40e6-acee-218947ab9e19-config\") pod \"ovsdbserver-nb-0\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.755398 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.755817 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9d816754-2d4c-40e6-acee-218947ab9e19-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.757198 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d816754-2d4c-40e6-acee-218947ab9e19-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.758387 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d816754-2d4c-40e6-acee-218947ab9e19-config\") pod \"ovsdbserver-nb-0\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.765118 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d816754-2d4c-40e6-acee-218947ab9e19-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.765359 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d816754-2d4c-40e6-acee-218947ab9e19-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.769535 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d816754-2d4c-40e6-acee-218947ab9e19-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.778396 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7tqx\" (UniqueName: \"kubernetes.io/projected/9d816754-2d4c-40e6-acee-218947ab9e19-kube-api-access-s7tqx\") pod \"ovsdbserver-nb-0\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.778649 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:27 crc kubenswrapper[4934]: I1002 10:05:27.860546 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.275949 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.279090 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.283279 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.283456 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.283797 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.286114 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-t9mg4" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.289050 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.418055 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6061002c-7242-4d02-b868-6838b39f30fa-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.418280 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6061002c-7242-4d02-b868-6838b39f30fa-config\") pod \"ovsdbserver-sb-0\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.418352 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6061002c-7242-4d02-b868-6838b39f30fa-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.418386 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfkgn\" (UniqueName: \"kubernetes.io/projected/6061002c-7242-4d02-b868-6838b39f30fa-kube-api-access-rfkgn\") pod \"ovsdbserver-sb-0\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.418443 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.418513 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6061002c-7242-4d02-b868-6838b39f30fa-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.418565 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6061002c-7242-4d02-b868-6838b39f30fa-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.418675 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6061002c-7242-4d02-b868-6838b39f30fa-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.520414 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6061002c-7242-4d02-b868-6838b39f30fa-config\") pod \"ovsdbserver-sb-0\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.520470 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6061002c-7242-4d02-b868-6838b39f30fa-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.520492 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfkgn\" (UniqueName: \"kubernetes.io/projected/6061002c-7242-4d02-b868-6838b39f30fa-kube-api-access-rfkgn\") pod \"ovsdbserver-sb-0\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.520517 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.520543 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6061002c-7242-4d02-b868-6838b39f30fa-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.520566 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6061002c-7242-4d02-b868-6838b39f30fa-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.520607 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6061002c-7242-4d02-b868-6838b39f30fa-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.520664 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6061002c-7242-4d02-b868-6838b39f30fa-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.521364 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6061002c-7242-4d02-b868-6838b39f30fa-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.521683 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.522480 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6061002c-7242-4d02-b868-6838b39f30fa-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.522959 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6061002c-7242-4d02-b868-6838b39f30fa-config\") pod \"ovsdbserver-sb-0\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.526564 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6061002c-7242-4d02-b868-6838b39f30fa-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.527068 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6061002c-7242-4d02-b868-6838b39f30fa-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.537569 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6061002c-7242-4d02-b868-6838b39f30fa-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.538176 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfkgn\" (UniqueName: \"kubernetes.io/projected/6061002c-7242-4d02-b868-6838b39f30fa-kube-api-access-rfkgn\") pod \"ovsdbserver-sb-0\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.559346 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:30 crc kubenswrapper[4934]: I1002 10:05:30.609116 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:32 crc kubenswrapper[4934]: I1002 10:05:32.934640 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 10:05:33 crc kubenswrapper[4934]: W1002 10:05:33.395334 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d5014eb_503f_45ea_bb78_bc5978dbaa69.slice/crio-96cc98d752154db530e353bca1dfee39ff89568e21354a3d2231ca857d5192b0 WatchSource:0}: Error finding container 96cc98d752154db530e353bca1dfee39ff89568e21354a3d2231ca857d5192b0: Status 404 returned error can't find the container with id 96cc98d752154db530e353bca1dfee39ff89568e21354a3d2231ca857d5192b0 Oct 02 10:05:33 crc kubenswrapper[4934]: E1002 10:05:33.426819 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:88f213718cf92b20884afce5905e9a604786648e459a80f0c95b2d2205f6e055" Oct 02 10:05:33 crc kubenswrapper[4934]: E1002 10:05:33.427236 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:88f213718cf92b20884afce5905e9a604786648e459a80f0c95b2d2205f6e055,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bjgwg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7968455899-qq7k2_openstack(dc548c73-564f-4830-a6f1-faa1912d39e6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 10:05:33 crc kubenswrapper[4934]: E1002 10:05:33.429826 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7968455899-qq7k2" podUID="dc548c73-564f-4830-a6f1-faa1912d39e6" Oct 02 10:05:33 crc kubenswrapper[4934]: E1002 10:05:33.435131 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:88f213718cf92b20884afce5905e9a604786648e459a80f0c95b2d2205f6e055" Oct 02 10:05:33 crc kubenswrapper[4934]: E1002 10:05:33.435284 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:88f213718cf92b20884afce5905e9a604786648e459a80f0c95b2d2205f6e055,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s85s2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-574db999b5-fw6n9_openstack(b952bbc6-20b0-45b3-90ce-04e7ef92a79c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 10:05:33 crc kubenswrapper[4934]: E1002 10:05:33.436469 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-574db999b5-fw6n9" podUID="b952bbc6-20b0-45b3-90ce-04e7ef92a79c" Oct 02 10:05:33 crc kubenswrapper[4934]: E1002 10:05:33.440481 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:88f213718cf92b20884afce5905e9a604786648e459a80f0c95b2d2205f6e055" Oct 02 10:05:33 crc kubenswrapper[4934]: E1002 10:05:33.440625 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:88f213718cf92b20884afce5905e9a604786648e459a80f0c95b2d2205f6e055,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-knlzs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6b5fff88f5-9wtdr_openstack(d8d5c788-c201-4daa-9b3c-3ea12cf696ff): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 10:05:33 crc kubenswrapper[4934]: E1002 10:05:33.442008 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-6b5fff88f5-9wtdr" podUID="d8d5c788-c201-4daa-9b3c-3ea12cf696ff" Oct 02 10:05:33 crc kubenswrapper[4934]: E1002 10:05:33.482452 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:88f213718cf92b20884afce5905e9a604786648e459a80f0c95b2d2205f6e055" Oct 02 10:05:33 crc kubenswrapper[4934]: E1002 10:05:33.482652 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:88f213718cf92b20884afce5905e9a604786648e459a80f0c95b2d2205f6e055,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b4hrl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7878bd5965-n75cl_openstack(a1ed141e-7d82-47c6-8d17-c1a1e1d601a7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 10:05:33 crc kubenswrapper[4934]: E1002 10:05:33.484739 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7878bd5965-n75cl" podUID="a1ed141e-7d82-47c6-8d17-c1a1e1d601a7" Oct 02 10:05:33 crc kubenswrapper[4934]: I1002 10:05:33.802728 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 02 10:05:33 crc kubenswrapper[4934]: W1002 10:05:33.804521 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c9f23b6_92f8_4d0a_8595_84c142adc3dd.slice/crio-612042673a7dc418a44963a83eed50002298c7097e79d4ad1bed5c687e90ef09 WatchSource:0}: Error finding container 612042673a7dc418a44963a83eed50002298c7097e79d4ad1bed5c687e90ef09: Status 404 returned error can't find the container with id 612042673a7dc418a44963a83eed50002298c7097e79d4ad1bed5c687e90ef09 Oct 02 10:05:33 crc kubenswrapper[4934]: I1002 10:05:33.960673 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 02 10:05:33 crc kubenswrapper[4934]: I1002 10:05:33.974794 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gwpkm"] Oct 02 10:05:33 crc kubenswrapper[4934]: W1002 10:05:33.975676 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7214291c_3c8e_4173_a6fd_ef16170258e2.slice/crio-5927bcbc36ec9b1bb775b2c5dde737fd22af01358495cd76bafa0b0d039a25cc WatchSource:0}: Error finding container 5927bcbc36ec9b1bb775b2c5dde737fd22af01358495cd76bafa0b0d039a25cc: Status 404 returned error can't find the container with id 5927bcbc36ec9b1bb775b2c5dde737fd22af01358495cd76bafa0b0d039a25cc Oct 02 10:05:33 crc kubenswrapper[4934]: W1002 10:05:33.978858 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf43ed4d4_3637_4ef6_a914_3a751cba8b4e.slice/crio-0297f7f74c5f85cb4628fdc1ed774e069f7c70d754ad3ed734b372387fd7c1e2 WatchSource:0}: Error finding container 0297f7f74c5f85cb4628fdc1ed774e069f7c70d754ad3ed734b372387fd7c1e2: Status 404 returned error can't find the container with id 0297f7f74c5f85cb4628fdc1ed774e069f7c70d754ad3ed734b372387fd7c1e2 Oct 02 10:05:33 crc kubenswrapper[4934]: W1002 10:05:33.980336 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod125b40d7_516c_432f_ae51_703f2b18068d.slice/crio-88252b1154831e865b61e0dd5b582c526915e28c56e8dd89163117a6b8f226a8 WatchSource:0}: Error finding container 88252b1154831e865b61e0dd5b582c526915e28c56e8dd89163117a6b8f226a8: Status 404 returned error can't find the container with id 88252b1154831e865b61e0dd5b582c526915e28c56e8dd89163117a6b8f226a8 Oct 02 10:05:33 crc kubenswrapper[4934]: I1002 10:05:33.989488 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 10:05:34 crc kubenswrapper[4934]: I1002 10:05:34.094792 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-65wtc"] Oct 02 10:05:34 crc kubenswrapper[4934]: I1002 10:05:34.112953 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gwpkm" event={"ID":"7214291c-3c8e-4173-a6fd-ef16170258e2","Type":"ContainerStarted","Data":"5927bcbc36ec9b1bb775b2c5dde737fd22af01358495cd76bafa0b0d039a25cc"} Oct 02 10:05:34 crc kubenswrapper[4934]: I1002 10:05:34.114546 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"f43ed4d4-3637-4ef6-a914-3a751cba8b4e","Type":"ContainerStarted","Data":"0297f7f74c5f85cb4628fdc1ed774e069f7c70d754ad3ed734b372387fd7c1e2"} Oct 02 10:05:34 crc kubenswrapper[4934]: I1002 10:05:34.116196 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"125b40d7-516c-432f-ae51-703f2b18068d","Type":"ContainerStarted","Data":"88252b1154831e865b61e0dd5b582c526915e28c56e8dd89163117a6b8f226a8"} Oct 02 10:05:34 crc kubenswrapper[4934]: I1002 10:05:34.117395 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8c9f23b6-92f8-4d0a-8595-84c142adc3dd","Type":"ContainerStarted","Data":"612042673a7dc418a44963a83eed50002298c7097e79d4ad1bed5c687e90ef09"} Oct 02 10:05:34 crc kubenswrapper[4934]: I1002 10:05:34.119129 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"0d5014eb-503f-45ea-bb78-bc5978dbaa69","Type":"ContainerStarted","Data":"96cc98d752154db530e353bca1dfee39ff89568e21354a3d2231ca857d5192b0"} Oct 02 10:05:34 crc kubenswrapper[4934]: E1002 10:05:34.122853 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:88f213718cf92b20884afce5905e9a604786648e459a80f0c95b2d2205f6e055\\\"\"" pod="openstack/dnsmasq-dns-6b5fff88f5-9wtdr" podUID="d8d5c788-c201-4daa-9b3c-3ea12cf696ff" Oct 02 10:05:34 crc kubenswrapper[4934]: E1002 10:05:34.123039 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:88f213718cf92b20884afce5905e9a604786648e459a80f0c95b2d2205f6e055\\\"\"" pod="openstack/dnsmasq-dns-7878bd5965-n75cl" podUID="a1ed141e-7d82-47c6-8d17-c1a1e1d601a7" Oct 02 10:05:34 crc kubenswrapper[4934]: I1002 10:05:34.207280 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 10:05:34 crc kubenswrapper[4934]: I1002 10:05:34.276021 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 10:05:34 crc kubenswrapper[4934]: W1002 10:05:34.291727 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6061002c_7242_4d02_b868_6838b39f30fa.slice/crio-cd21fda52b3b651db6f76212880c3304577978cd48904a4a334ce44fe9c84cd5 WatchSource:0}: Error finding container cd21fda52b3b651db6f76212880c3304577978cd48904a4a334ce44fe9c84cd5: Status 404 returned error can't find the container with id cd21fda52b3b651db6f76212880c3304577978cd48904a4a334ce44fe9c84cd5 Oct 02 10:05:34 crc kubenswrapper[4934]: W1002 10:05:34.292590 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d816754_2d4c_40e6_acee_218947ab9e19.slice/crio-3f2ad5d6f4972b1e2bcb07aa5ec6d1f71a859bef238a2c99f2a01002f7a2267b WatchSource:0}: Error finding container 3f2ad5d6f4972b1e2bcb07aa5ec6d1f71a859bef238a2c99f2a01002f7a2267b: Status 404 returned error can't find the container with id 3f2ad5d6f4972b1e2bcb07aa5ec6d1f71a859bef238a2c99f2a01002f7a2267b Oct 02 10:05:34 crc kubenswrapper[4934]: I1002 10:05:34.830978 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7968455899-qq7k2" Oct 02 10:05:34 crc kubenswrapper[4934]: I1002 10:05:34.846326 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-574db999b5-fw6n9" Oct 02 10:05:34 crc kubenswrapper[4934]: I1002 10:05:34.991968 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s85s2\" (UniqueName: \"kubernetes.io/projected/b952bbc6-20b0-45b3-90ce-04e7ef92a79c-kube-api-access-s85s2\") pod \"b952bbc6-20b0-45b3-90ce-04e7ef92a79c\" (UID: \"b952bbc6-20b0-45b3-90ce-04e7ef92a79c\") " Oct 02 10:05:34 crc kubenswrapper[4934]: I1002 10:05:34.992010 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc548c73-564f-4830-a6f1-faa1912d39e6-config\") pod \"dc548c73-564f-4830-a6f1-faa1912d39e6\" (UID: \"dc548c73-564f-4830-a6f1-faa1912d39e6\") " Oct 02 10:05:34 crc kubenswrapper[4934]: I1002 10:05:34.992036 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b952bbc6-20b0-45b3-90ce-04e7ef92a79c-config\") pod \"b952bbc6-20b0-45b3-90ce-04e7ef92a79c\" (UID: \"b952bbc6-20b0-45b3-90ce-04e7ef92a79c\") " Oct 02 10:05:34 crc kubenswrapper[4934]: I1002 10:05:34.992137 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjgwg\" (UniqueName: \"kubernetes.io/projected/dc548c73-564f-4830-a6f1-faa1912d39e6-kube-api-access-bjgwg\") pod \"dc548c73-564f-4830-a6f1-faa1912d39e6\" (UID: \"dc548c73-564f-4830-a6f1-faa1912d39e6\") " Oct 02 10:05:34 crc kubenswrapper[4934]: I1002 10:05:34.992185 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b952bbc6-20b0-45b3-90ce-04e7ef92a79c-dns-svc\") pod \"b952bbc6-20b0-45b3-90ce-04e7ef92a79c\" (UID: \"b952bbc6-20b0-45b3-90ce-04e7ef92a79c\") " Oct 02 10:05:34 crc kubenswrapper[4934]: I1002 10:05:34.992657 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b952bbc6-20b0-45b3-90ce-04e7ef92a79c-config" (OuterVolumeSpecName: "config") pod "b952bbc6-20b0-45b3-90ce-04e7ef92a79c" (UID: "b952bbc6-20b0-45b3-90ce-04e7ef92a79c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:05:34 crc kubenswrapper[4934]: I1002 10:05:34.992749 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b952bbc6-20b0-45b3-90ce-04e7ef92a79c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b952bbc6-20b0-45b3-90ce-04e7ef92a79c" (UID: "b952bbc6-20b0-45b3-90ce-04e7ef92a79c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:05:34 crc kubenswrapper[4934]: I1002 10:05:34.993270 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc548c73-564f-4830-a6f1-faa1912d39e6-config" (OuterVolumeSpecName: "config") pod "dc548c73-564f-4830-a6f1-faa1912d39e6" (UID: "dc548c73-564f-4830-a6f1-faa1912d39e6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:05:34 crc kubenswrapper[4934]: I1002 10:05:34.999802 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b952bbc6-20b0-45b3-90ce-04e7ef92a79c-kube-api-access-s85s2" (OuterVolumeSpecName: "kube-api-access-s85s2") pod "b952bbc6-20b0-45b3-90ce-04e7ef92a79c" (UID: "b952bbc6-20b0-45b3-90ce-04e7ef92a79c"). InnerVolumeSpecName "kube-api-access-s85s2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:05:35 crc kubenswrapper[4934]: I1002 10:05:35.003536 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc548c73-564f-4830-a6f1-faa1912d39e6-kube-api-access-bjgwg" (OuterVolumeSpecName: "kube-api-access-bjgwg") pod "dc548c73-564f-4830-a6f1-faa1912d39e6" (UID: "dc548c73-564f-4830-a6f1-faa1912d39e6"). InnerVolumeSpecName "kube-api-access-bjgwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:05:35 crc kubenswrapper[4934]: I1002 10:05:35.093570 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b952bbc6-20b0-45b3-90ce-04e7ef92a79c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:35 crc kubenswrapper[4934]: I1002 10:05:35.093616 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s85s2\" (UniqueName: \"kubernetes.io/projected/b952bbc6-20b0-45b3-90ce-04e7ef92a79c-kube-api-access-s85s2\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:35 crc kubenswrapper[4934]: I1002 10:05:35.093627 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc548c73-564f-4830-a6f1-faa1912d39e6-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:35 crc kubenswrapper[4934]: I1002 10:05:35.093636 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b952bbc6-20b0-45b3-90ce-04e7ef92a79c-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:35 crc kubenswrapper[4934]: I1002 10:05:35.093647 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjgwg\" (UniqueName: \"kubernetes.io/projected/dc548c73-564f-4830-a6f1-faa1912d39e6-kube-api-access-bjgwg\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:35 crc kubenswrapper[4934]: I1002 10:05:35.127629 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7968455899-qq7k2" event={"ID":"dc548c73-564f-4830-a6f1-faa1912d39e6","Type":"ContainerDied","Data":"11fbbbe18c5a52b50153581f88ee9726b0be8f47698ce3e8ca4747b60f8c9e9f"} Oct 02 10:05:35 crc kubenswrapper[4934]: I1002 10:05:35.127737 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7968455899-qq7k2" Oct 02 10:05:35 crc kubenswrapper[4934]: I1002 10:05:35.135796 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-65wtc" event={"ID":"1146cda2-78fc-40c5-b0d2-f8c19368fa2c","Type":"ContainerStarted","Data":"65bf9dacd49ccb2078166d218e8e99159a5e3b951f1f7cdf8a5023d8b0987280"} Oct 02 10:05:35 crc kubenswrapper[4934]: I1002 10:05:35.137870 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d2b6e083-a523-48e2-8157-088036174e12","Type":"ContainerStarted","Data":"eaa7cefa7a244d7ebc3161825f4958eb21752c3953ea843f3cde9449733066de"} Oct 02 10:05:35 crc kubenswrapper[4934]: I1002 10:05:35.139376 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"9d816754-2d4c-40e6-acee-218947ab9e19","Type":"ContainerStarted","Data":"3f2ad5d6f4972b1e2bcb07aa5ec6d1f71a859bef238a2c99f2a01002f7a2267b"} Oct 02 10:05:35 crc kubenswrapper[4934]: I1002 10:05:35.140748 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6061002c-7242-4d02-b868-6838b39f30fa","Type":"ContainerStarted","Data":"cd21fda52b3b651db6f76212880c3304577978cd48904a4a334ce44fe9c84cd5"} Oct 02 10:05:35 crc kubenswrapper[4934]: I1002 10:05:35.142330 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-574db999b5-fw6n9" Oct 02 10:05:35 crc kubenswrapper[4934]: I1002 10:05:35.142336 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-574db999b5-fw6n9" event={"ID":"b952bbc6-20b0-45b3-90ce-04e7ef92a79c","Type":"ContainerDied","Data":"1869ec405c80fde61d2320a26a2c0e850449f7d7319c87c6d30b367b10f16120"} Oct 02 10:05:35 crc kubenswrapper[4934]: I1002 10:05:35.143781 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"97560469-c459-48f4-901e-54eaded32537","Type":"ContainerStarted","Data":"91a99821c085c0ca3aac996ae178fce6d284a5d09eab41f0c4748d65f0d4756d"} Oct 02 10:05:35 crc kubenswrapper[4934]: I1002 10:05:35.206847 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7968455899-qq7k2"] Oct 02 10:05:35 crc kubenswrapper[4934]: I1002 10:05:35.214449 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7968455899-qq7k2"] Oct 02 10:05:35 crc kubenswrapper[4934]: I1002 10:05:35.231362 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-574db999b5-fw6n9"] Oct 02 10:05:35 crc kubenswrapper[4934]: I1002 10:05:35.236173 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-574db999b5-fw6n9"] Oct 02 10:05:36 crc kubenswrapper[4934]: I1002 10:05:36.926614 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b952bbc6-20b0-45b3-90ce-04e7ef92a79c" path="/var/lib/kubelet/pods/b952bbc6-20b0-45b3-90ce-04e7ef92a79c/volumes" Oct 02 10:05:36 crc kubenswrapper[4934]: I1002 10:05:36.927558 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc548c73-564f-4830-a6f1-faa1912d39e6" path="/var/lib/kubelet/pods/dc548c73-564f-4830-a6f1-faa1912d39e6/volumes" Oct 02 10:05:41 crc kubenswrapper[4934]: I1002 10:05:41.195206 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gwpkm" event={"ID":"7214291c-3c8e-4173-a6fd-ef16170258e2","Type":"ContainerStarted","Data":"5ab65d1f2a5fdc33711875776226a728389485f48a057c73387e2db56049cbbc"} Oct 02 10:05:41 crc kubenswrapper[4934]: I1002 10:05:41.195816 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-gwpkm" Oct 02 10:05:41 crc kubenswrapper[4934]: I1002 10:05:41.197897 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"f43ed4d4-3637-4ef6-a914-3a751cba8b4e","Type":"ContainerStarted","Data":"9d4708b7023ab963058c9d146e0e8bf04766c8e6b064f4e76425c8490490d0bd"} Oct 02 10:05:41 crc kubenswrapper[4934]: I1002 10:05:41.197989 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 02 10:05:41 crc kubenswrapper[4934]: I1002 10:05:41.200340 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"125b40d7-516c-432f-ae51-703f2b18068d","Type":"ContainerStarted","Data":"f6b1e12bbd4d9f68b287da7ba1aa9bfc9c5284fb40a0e164b3ec1825d7817335"} Oct 02 10:05:41 crc kubenswrapper[4934]: I1002 10:05:41.202656 4934 generic.go:334] "Generic (PLEG): container finished" podID="1146cda2-78fc-40c5-b0d2-f8c19368fa2c" containerID="73698957b5b31581f4624457e2c99dc25e8485557f6ad04a08edcb05551c73f3" exitCode=0 Oct 02 10:05:41 crc kubenswrapper[4934]: I1002 10:05:41.202725 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-65wtc" event={"ID":"1146cda2-78fc-40c5-b0d2-f8c19368fa2c","Type":"ContainerDied","Data":"73698957b5b31581f4624457e2c99dc25e8485557f6ad04a08edcb05551c73f3"} Oct 02 10:05:41 crc kubenswrapper[4934]: I1002 10:05:41.204979 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8c9f23b6-92f8-4d0a-8595-84c142adc3dd","Type":"ContainerStarted","Data":"f224fbd60ef61b586e97dd832e64710cef45baeff3c590ddb75d7e1d298abf45"} Oct 02 10:05:41 crc kubenswrapper[4934]: I1002 10:05:41.208049 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"9d816754-2d4c-40e6-acee-218947ab9e19","Type":"ContainerStarted","Data":"be30a7274c35eb6dd2db94c2805716e312b3406c00793273d691971991038504"} Oct 02 10:05:41 crc kubenswrapper[4934]: I1002 10:05:41.210769 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6061002c-7242-4d02-b868-6838b39f30fa","Type":"ContainerStarted","Data":"abc6a88d34ca1719e3b11b44524da713532a93344a46bb1626798816c0caa7a2"} Oct 02 10:05:41 crc kubenswrapper[4934]: I1002 10:05:41.213044 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"0d5014eb-503f-45ea-bb78-bc5978dbaa69","Type":"ContainerStarted","Data":"caccc0b41d120be902b372c1c34ce974b7a351bec8cbcf762fb7d2fb779776dd"} Oct 02 10:05:41 crc kubenswrapper[4934]: I1002 10:05:41.213119 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 02 10:05:41 crc kubenswrapper[4934]: I1002 10:05:41.231610 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-gwpkm" podStartSLOduration=8.814965969 podStartE2EDuration="15.231572612s" podCreationTimestamp="2025-10-02 10:05:26 +0000 UTC" firstStartedPulling="2025-10-02 10:05:33.98279111 +0000 UTC m=+1005.735432632" lastFinishedPulling="2025-10-02 10:05:40.399397753 +0000 UTC m=+1012.152039275" observedRunningTime="2025-10-02 10:05:41.224787158 +0000 UTC m=+1012.977428700" watchObservedRunningTime="2025-10-02 10:05:41.231572612 +0000 UTC m=+1012.984214134" Oct 02 10:05:41 crc kubenswrapper[4934]: I1002 10:05:41.309564 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=12.334785753 podStartE2EDuration="19.309541014s" podCreationTimestamp="2025-10-02 10:05:22 +0000 UTC" firstStartedPulling="2025-10-02 10:05:33.402695086 +0000 UTC m=+1005.155336628" lastFinishedPulling="2025-10-02 10:05:40.377450367 +0000 UTC m=+1012.130091889" observedRunningTime="2025-10-02 10:05:41.297850111 +0000 UTC m=+1013.050491653" watchObservedRunningTime="2025-10-02 10:05:41.309541014 +0000 UTC m=+1013.062182536" Oct 02 10:05:41 crc kubenswrapper[4934]: I1002 10:05:41.322511 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=13.94167411 podStartE2EDuration="20.322492513s" podCreationTimestamp="2025-10-02 10:05:21 +0000 UTC" firstStartedPulling="2025-10-02 10:05:33.983526831 +0000 UTC m=+1005.736168353" lastFinishedPulling="2025-10-02 10:05:40.364345194 +0000 UTC m=+1012.116986756" observedRunningTime="2025-10-02 10:05:41.314244718 +0000 UTC m=+1013.066886240" watchObservedRunningTime="2025-10-02 10:05:41.322492513 +0000 UTC m=+1013.075134045" Oct 02 10:05:42 crc kubenswrapper[4934]: I1002 10:05:42.230448 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-65wtc" event={"ID":"1146cda2-78fc-40c5-b0d2-f8c19368fa2c","Type":"ContainerStarted","Data":"c63d866adb61766a772229b32b348b743c5300884ef0ae4924c680c6cb85c6e0"} Oct 02 10:05:42 crc kubenswrapper[4934]: I1002 10:05:42.230799 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-65wtc" event={"ID":"1146cda2-78fc-40c5-b0d2-f8c19368fa2c","Type":"ContainerStarted","Data":"9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892"} Oct 02 10:05:43 crc kubenswrapper[4934]: I1002 10:05:43.247985 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-65wtc" Oct 02 10:05:43 crc kubenswrapper[4934]: I1002 10:05:43.248306 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-65wtc" Oct 02 10:05:44 crc kubenswrapper[4934]: I1002 10:05:44.258432 4934 generic.go:334] "Generic (PLEG): container finished" podID="8c9f23b6-92f8-4d0a-8595-84c142adc3dd" containerID="f224fbd60ef61b586e97dd832e64710cef45baeff3c590ddb75d7e1d298abf45" exitCode=0 Oct 02 10:05:44 crc kubenswrapper[4934]: I1002 10:05:44.258528 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8c9f23b6-92f8-4d0a-8595-84c142adc3dd","Type":"ContainerDied","Data":"f224fbd60ef61b586e97dd832e64710cef45baeff3c590ddb75d7e1d298abf45"} Oct 02 10:05:44 crc kubenswrapper[4934]: I1002 10:05:44.263507 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"9d816754-2d4c-40e6-acee-218947ab9e19","Type":"ContainerStarted","Data":"66355c51e4027fd1c4a3713d064cb84130067b07985283b47ad3d5d705b2d508"} Oct 02 10:05:44 crc kubenswrapper[4934]: I1002 10:05:44.267795 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6061002c-7242-4d02-b868-6838b39f30fa","Type":"ContainerStarted","Data":"fd9f74a1d204e8b395d9002c1eb07158525244b64cff40fb8c86051204dcedaa"} Oct 02 10:05:44 crc kubenswrapper[4934]: I1002 10:05:44.274883 4934 generic.go:334] "Generic (PLEG): container finished" podID="125b40d7-516c-432f-ae51-703f2b18068d" containerID="f6b1e12bbd4d9f68b287da7ba1aa9bfc9c5284fb40a0e164b3ec1825d7817335" exitCode=0 Oct 02 10:05:44 crc kubenswrapper[4934]: I1002 10:05:44.274963 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"125b40d7-516c-432f-ae51-703f2b18068d","Type":"ContainerDied","Data":"f6b1e12bbd4d9f68b287da7ba1aa9bfc9c5284fb40a0e164b3ec1825d7817335"} Oct 02 10:05:44 crc kubenswrapper[4934]: I1002 10:05:44.305250 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-65wtc" podStartSLOduration=12.131326496 podStartE2EDuration="18.305225931s" podCreationTimestamp="2025-10-02 10:05:26 +0000 UTC" firstStartedPulling="2025-10-02 10:05:34.212909439 +0000 UTC m=+1005.965550961" lastFinishedPulling="2025-10-02 10:05:40.386808874 +0000 UTC m=+1012.139450396" observedRunningTime="2025-10-02 10:05:42.253798369 +0000 UTC m=+1014.006439901" watchObservedRunningTime="2025-10-02 10:05:44.305225931 +0000 UTC m=+1016.057867493" Oct 02 10:05:44 crc kubenswrapper[4934]: I1002 10:05:44.364253 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=9.095779224 podStartE2EDuration="18.364235323s" podCreationTimestamp="2025-10-02 10:05:26 +0000 UTC" firstStartedPulling="2025-10-02 10:05:34.297458629 +0000 UTC m=+1006.050100151" lastFinishedPulling="2025-10-02 10:05:43.565914728 +0000 UTC m=+1015.318556250" observedRunningTime="2025-10-02 10:05:44.359524908 +0000 UTC m=+1016.112166440" watchObservedRunningTime="2025-10-02 10:05:44.364235323 +0000 UTC m=+1016.116876855" Oct 02 10:05:44 crc kubenswrapper[4934]: I1002 10:05:44.386530 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=6.098297566 podStartE2EDuration="15.386509748s" podCreationTimestamp="2025-10-02 10:05:29 +0000 UTC" firstStartedPulling="2025-10-02 10:05:34.294185946 +0000 UTC m=+1006.046827468" lastFinishedPulling="2025-10-02 10:05:43.582398128 +0000 UTC m=+1015.335039650" observedRunningTime="2025-10-02 10:05:44.379388454 +0000 UTC m=+1016.132030016" watchObservedRunningTime="2025-10-02 10:05:44.386509748 +0000 UTC m=+1016.139151280" Oct 02 10:05:45 crc kubenswrapper[4934]: I1002 10:05:45.288835 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8c9f23b6-92f8-4d0a-8595-84c142adc3dd","Type":"ContainerStarted","Data":"6ed60ff6b58b2bf5ef3faeb5193e2b929c4e3c03eaae93658e8fc29218880ae3"} Oct 02 10:05:45 crc kubenswrapper[4934]: I1002 10:05:45.292286 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"125b40d7-516c-432f-ae51-703f2b18068d","Type":"ContainerStarted","Data":"295f3add8c46cf850cd886f2d08991a6be7f2acfd564e68d63ce13cc92ba58df"} Oct 02 10:05:45 crc kubenswrapper[4934]: I1002 10:05:45.330349 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=19.746656855 podStartE2EDuration="26.330323689s" podCreationTimestamp="2025-10-02 10:05:19 +0000 UTC" firstStartedPulling="2025-10-02 10:05:33.806747472 +0000 UTC m=+1005.559388994" lastFinishedPulling="2025-10-02 10:05:40.390414306 +0000 UTC m=+1012.143055828" observedRunningTime="2025-10-02 10:05:45.321016423 +0000 UTC m=+1017.073657995" watchObservedRunningTime="2025-10-02 10:05:45.330323689 +0000 UTC m=+1017.082965241" Oct 02 10:05:45 crc kubenswrapper[4934]: I1002 10:05:45.355199 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=19.92753952 podStartE2EDuration="26.355176928s" podCreationTimestamp="2025-10-02 10:05:19 +0000 UTC" firstStartedPulling="2025-10-02 10:05:33.983325425 +0000 UTC m=+1005.735966947" lastFinishedPulling="2025-10-02 10:05:40.410962833 +0000 UTC m=+1012.163604355" observedRunningTime="2025-10-02 10:05:45.351154152 +0000 UTC m=+1017.103795744" watchObservedRunningTime="2025-10-02 10:05:45.355176928 +0000 UTC m=+1017.107818460" Oct 02 10:05:45 crc kubenswrapper[4934]: I1002 10:05:45.609649 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:45 crc kubenswrapper[4934]: I1002 10:05:45.610088 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:45 crc kubenswrapper[4934]: I1002 10:05:45.671430 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:45 crc kubenswrapper[4934]: I1002 10:05:45.861798 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:45 crc kubenswrapper[4934]: I1002 10:05:45.909967 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.304995 4934 generic.go:334] "Generic (PLEG): container finished" podID="d8d5c788-c201-4daa-9b3c-3ea12cf696ff" containerID="8d49c96ce7cdcce2add8644c0cd7f83e5d8576e97ea86ade9200b76b053d81e2" exitCode=0 Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.305113 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b5fff88f5-9wtdr" event={"ID":"d8d5c788-c201-4daa-9b3c-3ea12cf696ff","Type":"ContainerDied","Data":"8d49c96ce7cdcce2add8644c0cd7f83e5d8576e97ea86ade9200b76b053d81e2"} Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.305734 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.371630 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.383280 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.429719 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.677339 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b5fff88f5-9wtdr"] Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.733372 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-769b765785-hsdtk"] Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.734611 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-769b765785-hsdtk" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.741303 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.749967 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-769b765785-hsdtk"] Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.793345 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-l476l"] Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.794734 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-l476l" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.802402 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.804966 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-l476l"] Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.808021 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0302150f-a71a-45f0-a6b4-024664b21e51-ovsdbserver-nb\") pod \"dnsmasq-dns-769b765785-hsdtk\" (UID: \"0302150f-a71a-45f0-a6b4-024664b21e51\") " pod="openstack/dnsmasq-dns-769b765785-hsdtk" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.808080 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ls99k\" (UniqueName: \"kubernetes.io/projected/0302150f-a71a-45f0-a6b4-024664b21e51-kube-api-access-ls99k\") pod \"dnsmasq-dns-769b765785-hsdtk\" (UID: \"0302150f-a71a-45f0-a6b4-024664b21e51\") " pod="openstack/dnsmasq-dns-769b765785-hsdtk" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.808114 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0302150f-a71a-45f0-a6b4-024664b21e51-config\") pod \"dnsmasq-dns-769b765785-hsdtk\" (UID: \"0302150f-a71a-45f0-a6b4-024664b21e51\") " pod="openstack/dnsmasq-dns-769b765785-hsdtk" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.808175 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0302150f-a71a-45f0-a6b4-024664b21e51-dns-svc\") pod \"dnsmasq-dns-769b765785-hsdtk\" (UID: \"0302150f-a71a-45f0-a6b4-024664b21e51\") " pod="openstack/dnsmasq-dns-769b765785-hsdtk" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.861497 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7878bd5965-n75cl"] Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.872865 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.874257 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.875819 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.876234 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.876467 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.876637 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-97nm7" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.900394 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.913501 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvqfn\" (UniqueName: \"kubernetes.io/projected/e14738be-f83f-47ce-b0d9-ceb23d504b46-kube-api-access-mvqfn\") pod \"ovn-controller-metrics-l476l\" (UID: \"e14738be-f83f-47ce-b0d9-ceb23d504b46\") " pod="openstack/ovn-controller-metrics-l476l" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.913548 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e14738be-f83f-47ce-b0d9-ceb23d504b46-ovs-rundir\") pod \"ovn-controller-metrics-l476l\" (UID: \"e14738be-f83f-47ce-b0d9-ceb23d504b46\") " pod="openstack/ovn-controller-metrics-l476l" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.913710 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e14738be-f83f-47ce-b0d9-ceb23d504b46-ovn-rundir\") pod \"ovn-controller-metrics-l476l\" (UID: \"e14738be-f83f-47ce-b0d9-ceb23d504b46\") " pod="openstack/ovn-controller-metrics-l476l" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.913734 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e14738be-f83f-47ce-b0d9-ceb23d504b46-config\") pod \"ovn-controller-metrics-l476l\" (UID: \"e14738be-f83f-47ce-b0d9-ceb23d504b46\") " pod="openstack/ovn-controller-metrics-l476l" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.913757 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0302150f-a71a-45f0-a6b4-024664b21e51-dns-svc\") pod \"dnsmasq-dns-769b765785-hsdtk\" (UID: \"0302150f-a71a-45f0-a6b4-024664b21e51\") " pod="openstack/dnsmasq-dns-769b765785-hsdtk" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.913790 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e14738be-f83f-47ce-b0d9-ceb23d504b46-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-l476l\" (UID: \"e14738be-f83f-47ce-b0d9-ceb23d504b46\") " pod="openstack/ovn-controller-metrics-l476l" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.913831 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0302150f-a71a-45f0-a6b4-024664b21e51-ovsdbserver-nb\") pod \"dnsmasq-dns-769b765785-hsdtk\" (UID: \"0302150f-a71a-45f0-a6b4-024664b21e51\") " pod="openstack/dnsmasq-dns-769b765785-hsdtk" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.913864 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ls99k\" (UniqueName: \"kubernetes.io/projected/0302150f-a71a-45f0-a6b4-024664b21e51-kube-api-access-ls99k\") pod \"dnsmasq-dns-769b765785-hsdtk\" (UID: \"0302150f-a71a-45f0-a6b4-024664b21e51\") " pod="openstack/dnsmasq-dns-769b765785-hsdtk" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.913893 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0302150f-a71a-45f0-a6b4-024664b21e51-config\") pod \"dnsmasq-dns-769b765785-hsdtk\" (UID: \"0302150f-a71a-45f0-a6b4-024664b21e51\") " pod="openstack/dnsmasq-dns-769b765785-hsdtk" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.913913 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e14738be-f83f-47ce-b0d9-ceb23d504b46-combined-ca-bundle\") pod \"ovn-controller-metrics-l476l\" (UID: \"e14738be-f83f-47ce-b0d9-ceb23d504b46\") " pod="openstack/ovn-controller-metrics-l476l" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.914883 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0302150f-a71a-45f0-a6b4-024664b21e51-config\") pod \"dnsmasq-dns-769b765785-hsdtk\" (UID: \"0302150f-a71a-45f0-a6b4-024664b21e51\") " pod="openstack/dnsmasq-dns-769b765785-hsdtk" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.914895 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0302150f-a71a-45f0-a6b4-024664b21e51-ovsdbserver-nb\") pod \"dnsmasq-dns-769b765785-hsdtk\" (UID: \"0302150f-a71a-45f0-a6b4-024664b21e51\") " pod="openstack/dnsmasq-dns-769b765785-hsdtk" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.915525 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0302150f-a71a-45f0-a6b4-024664b21e51-dns-svc\") pod \"dnsmasq-dns-769b765785-hsdtk\" (UID: \"0302150f-a71a-45f0-a6b4-024664b21e51\") " pod="openstack/dnsmasq-dns-769b765785-hsdtk" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.929712 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bd4b77fc7-w8tlj"] Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.937340 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.952148 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ls99k\" (UniqueName: \"kubernetes.io/projected/0302150f-a71a-45f0-a6b4-024664b21e51-kube-api-access-ls99k\") pod \"dnsmasq-dns-769b765785-hsdtk\" (UID: \"0302150f-a71a-45f0-a6b4-024664b21e51\") " pod="openstack/dnsmasq-dns-769b765785-hsdtk" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.954043 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 02 10:05:46 crc kubenswrapper[4934]: I1002 10:05:46.982943 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bd4b77fc7-w8tlj"] Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.015027 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/05c8faa0-6986-415d-abc3-4d35291375f2-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"05c8faa0-6986-415d-abc3-4d35291375f2\") " pod="openstack/ovn-northd-0" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.015077 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05c8faa0-6986-415d-abc3-4d35291375f2-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"05c8faa0-6986-415d-abc3-4d35291375f2\") " pod="openstack/ovn-northd-0" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.015103 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05c8faa0-6986-415d-abc3-4d35291375f2-config\") pod \"ovn-northd-0\" (UID: \"05c8faa0-6986-415d-abc3-4d35291375f2\") " pod="openstack/ovn-northd-0" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.015151 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3c7b1c5-61ea-4573-8709-252ea669dd2b-dns-svc\") pod \"dnsmasq-dns-5bd4b77fc7-w8tlj\" (UID: \"b3c7b1c5-61ea-4573-8709-252ea669dd2b\") " pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.015262 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/05c8faa0-6986-415d-abc3-4d35291375f2-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"05c8faa0-6986-415d-abc3-4d35291375f2\") " pod="openstack/ovn-northd-0" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.015298 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/05c8faa0-6986-415d-abc3-4d35291375f2-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"05c8faa0-6986-415d-abc3-4d35291375f2\") " pod="openstack/ovn-northd-0" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.015317 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3c7b1c5-61ea-4573-8709-252ea669dd2b-config\") pod \"dnsmasq-dns-5bd4b77fc7-w8tlj\" (UID: \"b3c7b1c5-61ea-4573-8709-252ea669dd2b\") " pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.015347 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3c7b1c5-61ea-4573-8709-252ea669dd2b-ovsdbserver-sb\") pod \"dnsmasq-dns-5bd4b77fc7-w8tlj\" (UID: \"b3c7b1c5-61ea-4573-8709-252ea669dd2b\") " pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.015377 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e14738be-f83f-47ce-b0d9-ceb23d504b46-combined-ca-bundle\") pod \"ovn-controller-metrics-l476l\" (UID: \"e14738be-f83f-47ce-b0d9-ceb23d504b46\") " pod="openstack/ovn-controller-metrics-l476l" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.015457 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7496\" (UniqueName: \"kubernetes.io/projected/05c8faa0-6986-415d-abc3-4d35291375f2-kube-api-access-x7496\") pod \"ovn-northd-0\" (UID: \"05c8faa0-6986-415d-abc3-4d35291375f2\") " pod="openstack/ovn-northd-0" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.015503 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3c7b1c5-61ea-4573-8709-252ea669dd2b-ovsdbserver-nb\") pod \"dnsmasq-dns-5bd4b77fc7-w8tlj\" (UID: \"b3c7b1c5-61ea-4573-8709-252ea669dd2b\") " pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.015532 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldqdc\" (UniqueName: \"kubernetes.io/projected/b3c7b1c5-61ea-4573-8709-252ea669dd2b-kube-api-access-ldqdc\") pod \"dnsmasq-dns-5bd4b77fc7-w8tlj\" (UID: \"b3c7b1c5-61ea-4573-8709-252ea669dd2b\") " pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.015677 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvqfn\" (UniqueName: \"kubernetes.io/projected/e14738be-f83f-47ce-b0d9-ceb23d504b46-kube-api-access-mvqfn\") pod \"ovn-controller-metrics-l476l\" (UID: \"e14738be-f83f-47ce-b0d9-ceb23d504b46\") " pod="openstack/ovn-controller-metrics-l476l" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.015994 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e14738be-f83f-47ce-b0d9-ceb23d504b46-ovs-rundir\") pod \"ovn-controller-metrics-l476l\" (UID: \"e14738be-f83f-47ce-b0d9-ceb23d504b46\") " pod="openstack/ovn-controller-metrics-l476l" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.016051 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e14738be-f83f-47ce-b0d9-ceb23d504b46-ovn-rundir\") pod \"ovn-controller-metrics-l476l\" (UID: \"e14738be-f83f-47ce-b0d9-ceb23d504b46\") " pod="openstack/ovn-controller-metrics-l476l" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.016081 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e14738be-f83f-47ce-b0d9-ceb23d504b46-config\") pod \"ovn-controller-metrics-l476l\" (UID: \"e14738be-f83f-47ce-b0d9-ceb23d504b46\") " pod="openstack/ovn-controller-metrics-l476l" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.016122 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/05c8faa0-6986-415d-abc3-4d35291375f2-scripts\") pod \"ovn-northd-0\" (UID: \"05c8faa0-6986-415d-abc3-4d35291375f2\") " pod="openstack/ovn-northd-0" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.016161 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e14738be-f83f-47ce-b0d9-ceb23d504b46-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-l476l\" (UID: \"e14738be-f83f-47ce-b0d9-ceb23d504b46\") " pod="openstack/ovn-controller-metrics-l476l" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.016251 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e14738be-f83f-47ce-b0d9-ceb23d504b46-ovs-rundir\") pod \"ovn-controller-metrics-l476l\" (UID: \"e14738be-f83f-47ce-b0d9-ceb23d504b46\") " pod="openstack/ovn-controller-metrics-l476l" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.016324 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e14738be-f83f-47ce-b0d9-ceb23d504b46-ovn-rundir\") pod \"ovn-controller-metrics-l476l\" (UID: \"e14738be-f83f-47ce-b0d9-ceb23d504b46\") " pod="openstack/ovn-controller-metrics-l476l" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.016864 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e14738be-f83f-47ce-b0d9-ceb23d504b46-config\") pod \"ovn-controller-metrics-l476l\" (UID: \"e14738be-f83f-47ce-b0d9-ceb23d504b46\") " pod="openstack/ovn-controller-metrics-l476l" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.019602 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e14738be-f83f-47ce-b0d9-ceb23d504b46-combined-ca-bundle\") pod \"ovn-controller-metrics-l476l\" (UID: \"e14738be-f83f-47ce-b0d9-ceb23d504b46\") " pod="openstack/ovn-controller-metrics-l476l" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.021124 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e14738be-f83f-47ce-b0d9-ceb23d504b46-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-l476l\" (UID: \"e14738be-f83f-47ce-b0d9-ceb23d504b46\") " pod="openstack/ovn-controller-metrics-l476l" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.033554 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvqfn\" (UniqueName: \"kubernetes.io/projected/e14738be-f83f-47ce-b0d9-ceb23d504b46-kube-api-access-mvqfn\") pod \"ovn-controller-metrics-l476l\" (UID: \"e14738be-f83f-47ce-b0d9-ceb23d504b46\") " pod="openstack/ovn-controller-metrics-l476l" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.052343 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-769b765785-hsdtk" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.117466 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7496\" (UniqueName: \"kubernetes.io/projected/05c8faa0-6986-415d-abc3-4d35291375f2-kube-api-access-x7496\") pod \"ovn-northd-0\" (UID: \"05c8faa0-6986-415d-abc3-4d35291375f2\") " pod="openstack/ovn-northd-0" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.117516 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3c7b1c5-61ea-4573-8709-252ea669dd2b-ovsdbserver-nb\") pod \"dnsmasq-dns-5bd4b77fc7-w8tlj\" (UID: \"b3c7b1c5-61ea-4573-8709-252ea669dd2b\") " pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.117537 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldqdc\" (UniqueName: \"kubernetes.io/projected/b3c7b1c5-61ea-4573-8709-252ea669dd2b-kube-api-access-ldqdc\") pod \"dnsmasq-dns-5bd4b77fc7-w8tlj\" (UID: \"b3c7b1c5-61ea-4573-8709-252ea669dd2b\") " pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.117620 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/05c8faa0-6986-415d-abc3-4d35291375f2-scripts\") pod \"ovn-northd-0\" (UID: \"05c8faa0-6986-415d-abc3-4d35291375f2\") " pod="openstack/ovn-northd-0" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.117659 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/05c8faa0-6986-415d-abc3-4d35291375f2-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"05c8faa0-6986-415d-abc3-4d35291375f2\") " pod="openstack/ovn-northd-0" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.117683 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05c8faa0-6986-415d-abc3-4d35291375f2-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"05c8faa0-6986-415d-abc3-4d35291375f2\") " pod="openstack/ovn-northd-0" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.117703 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05c8faa0-6986-415d-abc3-4d35291375f2-config\") pod \"ovn-northd-0\" (UID: \"05c8faa0-6986-415d-abc3-4d35291375f2\") " pod="openstack/ovn-northd-0" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.117732 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3c7b1c5-61ea-4573-8709-252ea669dd2b-dns-svc\") pod \"dnsmasq-dns-5bd4b77fc7-w8tlj\" (UID: \"b3c7b1c5-61ea-4573-8709-252ea669dd2b\") " pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.117752 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/05c8faa0-6986-415d-abc3-4d35291375f2-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"05c8faa0-6986-415d-abc3-4d35291375f2\") " pod="openstack/ovn-northd-0" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.117775 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/05c8faa0-6986-415d-abc3-4d35291375f2-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"05c8faa0-6986-415d-abc3-4d35291375f2\") " pod="openstack/ovn-northd-0" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.117790 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3c7b1c5-61ea-4573-8709-252ea669dd2b-config\") pod \"dnsmasq-dns-5bd4b77fc7-w8tlj\" (UID: \"b3c7b1c5-61ea-4573-8709-252ea669dd2b\") " pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.117805 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3c7b1c5-61ea-4573-8709-252ea669dd2b-ovsdbserver-sb\") pod \"dnsmasq-dns-5bd4b77fc7-w8tlj\" (UID: \"b3c7b1c5-61ea-4573-8709-252ea669dd2b\") " pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.118618 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3c7b1c5-61ea-4573-8709-252ea669dd2b-ovsdbserver-nb\") pod \"dnsmasq-dns-5bd4b77fc7-w8tlj\" (UID: \"b3c7b1c5-61ea-4573-8709-252ea669dd2b\") " pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.118651 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/05c8faa0-6986-415d-abc3-4d35291375f2-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"05c8faa0-6986-415d-abc3-4d35291375f2\") " pod="openstack/ovn-northd-0" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.118687 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3c7b1c5-61ea-4573-8709-252ea669dd2b-ovsdbserver-sb\") pod \"dnsmasq-dns-5bd4b77fc7-w8tlj\" (UID: \"b3c7b1c5-61ea-4573-8709-252ea669dd2b\") " pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.118725 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05c8faa0-6986-415d-abc3-4d35291375f2-config\") pod \"ovn-northd-0\" (UID: \"05c8faa0-6986-415d-abc3-4d35291375f2\") " pod="openstack/ovn-northd-0" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.118750 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/05c8faa0-6986-415d-abc3-4d35291375f2-scripts\") pod \"ovn-northd-0\" (UID: \"05c8faa0-6986-415d-abc3-4d35291375f2\") " pod="openstack/ovn-northd-0" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.119372 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3c7b1c5-61ea-4573-8709-252ea669dd2b-dns-svc\") pod \"dnsmasq-dns-5bd4b77fc7-w8tlj\" (UID: \"b3c7b1c5-61ea-4573-8709-252ea669dd2b\") " pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.119598 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-l476l" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.120042 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3c7b1c5-61ea-4573-8709-252ea669dd2b-config\") pod \"dnsmasq-dns-5bd4b77fc7-w8tlj\" (UID: \"b3c7b1c5-61ea-4573-8709-252ea669dd2b\") " pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.122288 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05c8faa0-6986-415d-abc3-4d35291375f2-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"05c8faa0-6986-415d-abc3-4d35291375f2\") " pod="openstack/ovn-northd-0" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.122313 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/05c8faa0-6986-415d-abc3-4d35291375f2-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"05c8faa0-6986-415d-abc3-4d35291375f2\") " pod="openstack/ovn-northd-0" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.123703 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/05c8faa0-6986-415d-abc3-4d35291375f2-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"05c8faa0-6986-415d-abc3-4d35291375f2\") " pod="openstack/ovn-northd-0" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.140662 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7878bd5965-n75cl" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.141672 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7496\" (UniqueName: \"kubernetes.io/projected/05c8faa0-6986-415d-abc3-4d35291375f2-kube-api-access-x7496\") pod \"ovn-northd-0\" (UID: \"05c8faa0-6986-415d-abc3-4d35291375f2\") " pod="openstack/ovn-northd-0" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.147971 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldqdc\" (UniqueName: \"kubernetes.io/projected/b3c7b1c5-61ea-4573-8709-252ea669dd2b-kube-api-access-ldqdc\") pod \"dnsmasq-dns-5bd4b77fc7-w8tlj\" (UID: \"b3c7b1c5-61ea-4573-8709-252ea669dd2b\") " pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.201013 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.218986 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1ed141e-7d82-47c6-8d17-c1a1e1d601a7-dns-svc\") pod \"a1ed141e-7d82-47c6-8d17-c1a1e1d601a7\" (UID: \"a1ed141e-7d82-47c6-8d17-c1a1e1d601a7\") " Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.219050 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4hrl\" (UniqueName: \"kubernetes.io/projected/a1ed141e-7d82-47c6-8d17-c1a1e1d601a7-kube-api-access-b4hrl\") pod \"a1ed141e-7d82-47c6-8d17-c1a1e1d601a7\" (UID: \"a1ed141e-7d82-47c6-8d17-c1a1e1d601a7\") " Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.219168 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1ed141e-7d82-47c6-8d17-c1a1e1d601a7-config\") pod \"a1ed141e-7d82-47c6-8d17-c1a1e1d601a7\" (UID: \"a1ed141e-7d82-47c6-8d17-c1a1e1d601a7\") " Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.219851 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1ed141e-7d82-47c6-8d17-c1a1e1d601a7-config" (OuterVolumeSpecName: "config") pod "a1ed141e-7d82-47c6-8d17-c1a1e1d601a7" (UID: "a1ed141e-7d82-47c6-8d17-c1a1e1d601a7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.219875 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1ed141e-7d82-47c6-8d17-c1a1e1d601a7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a1ed141e-7d82-47c6-8d17-c1a1e1d601a7" (UID: "a1ed141e-7d82-47c6-8d17-c1a1e1d601a7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.222636 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1ed141e-7d82-47c6-8d17-c1a1e1d601a7-kube-api-access-b4hrl" (OuterVolumeSpecName: "kube-api-access-b4hrl") pod "a1ed141e-7d82-47c6-8d17-c1a1e1d601a7" (UID: "a1ed141e-7d82-47c6-8d17-c1a1e1d601a7"). InnerVolumeSpecName "kube-api-access-b4hrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.299169 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.319058 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7878bd5965-n75cl" event={"ID":"a1ed141e-7d82-47c6-8d17-c1a1e1d601a7","Type":"ContainerDied","Data":"a58c1061224f039e736f8aaef018231338d07eaa3725105d31a0338383520abc"} Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.319137 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7878bd5965-n75cl" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.320626 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1ed141e-7d82-47c6-8d17-c1a1e1d601a7-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.320649 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1ed141e-7d82-47c6-8d17-c1a1e1d601a7-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.320658 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4hrl\" (UniqueName: \"kubernetes.io/projected/a1ed141e-7d82-47c6-8d17-c1a1e1d601a7-kube-api-access-b4hrl\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.322203 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b5fff88f5-9wtdr" podUID="d8d5c788-c201-4daa-9b3c-3ea12cf696ff" containerName="dnsmasq-dns" containerID="cri-o://8d2579d409297d2b8abaf65ea5bb98ea1c241b6eda4c7c4f4d24d9d481f1a07c" gracePeriod=10 Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.322414 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b5fff88f5-9wtdr" event={"ID":"d8d5c788-c201-4daa-9b3c-3ea12cf696ff","Type":"ContainerStarted","Data":"8d2579d409297d2b8abaf65ea5bb98ea1c241b6eda4c7c4f4d24d9d481f1a07c"} Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.322912 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b5fff88f5-9wtdr" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.341272 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b5fff88f5-9wtdr" podStartSLOduration=3.624145725 podStartE2EDuration="31.341202445s" podCreationTimestamp="2025-10-02 10:05:16 +0000 UTC" firstStartedPulling="2025-10-02 10:05:17.641674619 +0000 UTC m=+989.394316141" lastFinishedPulling="2025-10-02 10:05:45.358731319 +0000 UTC m=+1017.111372861" observedRunningTime="2025-10-02 10:05:47.340460843 +0000 UTC m=+1019.093102375" watchObservedRunningTime="2025-10-02 10:05:47.341202445 +0000 UTC m=+1019.093843967" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.490375 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7878bd5965-n75cl"] Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.505699 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7878bd5965-n75cl"] Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.506039 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-769b765785-hsdtk"] Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.596263 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-l476l"] Oct 02 10:05:47 crc kubenswrapper[4934]: W1002 10:05:47.607106 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode14738be_f83f_47ce_b0d9_ceb23d504b46.slice/crio-02a4994e4aa0fe60ba80ae18379efa8ae4e5356c97b4752a007e79f341858e26 WatchSource:0}: Error finding container 02a4994e4aa0fe60ba80ae18379efa8ae4e5356c97b4752a007e79f341858e26: Status 404 returned error can't find the container with id 02a4994e4aa0fe60ba80ae18379efa8ae4e5356c97b4752a007e79f341858e26 Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.755104 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.801730 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bd4b77fc7-w8tlj"] Oct 02 10:05:47 crc kubenswrapper[4934]: W1002 10:05:47.843887 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3c7b1c5_61ea_4573_8709_252ea669dd2b.slice/crio-ee7cf74e3d31a22106b015efac21da25bde0e7f6d4ea31b2475764bfd094a780 WatchSource:0}: Error finding container ee7cf74e3d31a22106b015efac21da25bde0e7f6d4ea31b2475764bfd094a780: Status 404 returned error can't find the container with id ee7cf74e3d31a22106b015efac21da25bde0e7f6d4ea31b2475764bfd094a780 Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.846133 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b5fff88f5-9wtdr" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.934545 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8d5c788-c201-4daa-9b3c-3ea12cf696ff-dns-svc\") pod \"d8d5c788-c201-4daa-9b3c-3ea12cf696ff\" (UID: \"d8d5c788-c201-4daa-9b3c-3ea12cf696ff\") " Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.934983 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knlzs\" (UniqueName: \"kubernetes.io/projected/d8d5c788-c201-4daa-9b3c-3ea12cf696ff-kube-api-access-knlzs\") pod \"d8d5c788-c201-4daa-9b3c-3ea12cf696ff\" (UID: \"d8d5c788-c201-4daa-9b3c-3ea12cf696ff\") " Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.935094 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8d5c788-c201-4daa-9b3c-3ea12cf696ff-config\") pod \"d8d5c788-c201-4daa-9b3c-3ea12cf696ff\" (UID: \"d8d5c788-c201-4daa-9b3c-3ea12cf696ff\") " Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.940115 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8d5c788-c201-4daa-9b3c-3ea12cf696ff-kube-api-access-knlzs" (OuterVolumeSpecName: "kube-api-access-knlzs") pod "d8d5c788-c201-4daa-9b3c-3ea12cf696ff" (UID: "d8d5c788-c201-4daa-9b3c-3ea12cf696ff"). InnerVolumeSpecName "kube-api-access-knlzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.981416 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8d5c788-c201-4daa-9b3c-3ea12cf696ff-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d8d5c788-c201-4daa-9b3c-3ea12cf696ff" (UID: "d8d5c788-c201-4daa-9b3c-3ea12cf696ff"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:05:47 crc kubenswrapper[4934]: I1002 10:05:47.984773 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8d5c788-c201-4daa-9b3c-3ea12cf696ff-config" (OuterVolumeSpecName: "config") pod "d8d5c788-c201-4daa-9b3c-3ea12cf696ff" (UID: "d8d5c788-c201-4daa-9b3c-3ea12cf696ff"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:05:48 crc kubenswrapper[4934]: I1002 10:05:48.036973 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d8d5c788-c201-4daa-9b3c-3ea12cf696ff-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:48 crc kubenswrapper[4934]: I1002 10:05:48.037010 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knlzs\" (UniqueName: \"kubernetes.io/projected/d8d5c788-c201-4daa-9b3c-3ea12cf696ff-kube-api-access-knlzs\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:48 crc kubenswrapper[4934]: I1002 10:05:48.037024 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8d5c788-c201-4daa-9b3c-3ea12cf696ff-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:05:48 crc kubenswrapper[4934]: I1002 10:05:48.334155 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-l476l" event={"ID":"e14738be-f83f-47ce-b0d9-ceb23d504b46","Type":"ContainerStarted","Data":"336b71c58bd0e5981fd64408bac6c360461733df6dbdbba45a423f97c6077a34"} Oct 02 10:05:48 crc kubenswrapper[4934]: I1002 10:05:48.334409 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-l476l" event={"ID":"e14738be-f83f-47ce-b0d9-ceb23d504b46","Type":"ContainerStarted","Data":"02a4994e4aa0fe60ba80ae18379efa8ae4e5356c97b4752a007e79f341858e26"} Oct 02 10:05:48 crc kubenswrapper[4934]: I1002 10:05:48.338073 4934 generic.go:334] "Generic (PLEG): container finished" podID="b3c7b1c5-61ea-4573-8709-252ea669dd2b" containerID="f29bffa051f3ee598a217da3e18a3447a85c5230a3130c9f68e351e0b0e00403" exitCode=0 Oct 02 10:05:48 crc kubenswrapper[4934]: I1002 10:05:48.338112 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" event={"ID":"b3c7b1c5-61ea-4573-8709-252ea669dd2b","Type":"ContainerDied","Data":"f29bffa051f3ee598a217da3e18a3447a85c5230a3130c9f68e351e0b0e00403"} Oct 02 10:05:48 crc kubenswrapper[4934]: I1002 10:05:48.338125 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" event={"ID":"b3c7b1c5-61ea-4573-8709-252ea669dd2b","Type":"ContainerStarted","Data":"ee7cf74e3d31a22106b015efac21da25bde0e7f6d4ea31b2475764bfd094a780"} Oct 02 10:05:48 crc kubenswrapper[4934]: I1002 10:05:48.341054 4934 generic.go:334] "Generic (PLEG): container finished" podID="d8d5c788-c201-4daa-9b3c-3ea12cf696ff" containerID="8d2579d409297d2b8abaf65ea5bb98ea1c241b6eda4c7c4f4d24d9d481f1a07c" exitCode=0 Oct 02 10:05:48 crc kubenswrapper[4934]: I1002 10:05:48.341218 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b5fff88f5-9wtdr" Oct 02 10:05:48 crc kubenswrapper[4934]: I1002 10:05:48.341504 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b5fff88f5-9wtdr" event={"ID":"d8d5c788-c201-4daa-9b3c-3ea12cf696ff","Type":"ContainerDied","Data":"8d2579d409297d2b8abaf65ea5bb98ea1c241b6eda4c7c4f4d24d9d481f1a07c"} Oct 02 10:05:48 crc kubenswrapper[4934]: I1002 10:05:48.341553 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b5fff88f5-9wtdr" event={"ID":"d8d5c788-c201-4daa-9b3c-3ea12cf696ff","Type":"ContainerDied","Data":"09b782665b0760fdbfbdfd37dca5dc87592b65a07d1bd5b9d4141ac729c271a0"} Oct 02 10:05:48 crc kubenswrapper[4934]: I1002 10:05:48.341592 4934 scope.go:117] "RemoveContainer" containerID="8d2579d409297d2b8abaf65ea5bb98ea1c241b6eda4c7c4f4d24d9d481f1a07c" Oct 02 10:05:48 crc kubenswrapper[4934]: I1002 10:05:48.356644 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-l476l" podStartSLOduration=2.356624097 podStartE2EDuration="2.356624097s" podCreationTimestamp="2025-10-02 10:05:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:05:48.354343002 +0000 UTC m=+1020.106984564" watchObservedRunningTime="2025-10-02 10:05:48.356624097 +0000 UTC m=+1020.109265619" Oct 02 10:05:48 crc kubenswrapper[4934]: I1002 10:05:48.370782 4934 generic.go:334] "Generic (PLEG): container finished" podID="0302150f-a71a-45f0-a6b4-024664b21e51" containerID="d5bf4f453290c95a9ed1c0b73753ce7ef8c66e50a522b798669bfe817dfa15a3" exitCode=0 Oct 02 10:05:48 crc kubenswrapper[4934]: I1002 10:05:48.370899 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-769b765785-hsdtk" event={"ID":"0302150f-a71a-45f0-a6b4-024664b21e51","Type":"ContainerDied","Data":"d5bf4f453290c95a9ed1c0b73753ce7ef8c66e50a522b798669bfe817dfa15a3"} Oct 02 10:05:48 crc kubenswrapper[4934]: I1002 10:05:48.370929 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-769b765785-hsdtk" event={"ID":"0302150f-a71a-45f0-a6b4-024664b21e51","Type":"ContainerStarted","Data":"705ad62eab9f24bd290443018ac7fc5b141723a5b8ef9cec64a7e76dd0402e31"} Oct 02 10:05:48 crc kubenswrapper[4934]: I1002 10:05:48.385098 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"05c8faa0-6986-415d-abc3-4d35291375f2","Type":"ContainerStarted","Data":"582a5c1df61a40fae2fe36f824dd290361362b220d5f4f143379cc93d0cb24c5"} Oct 02 10:05:48 crc kubenswrapper[4934]: I1002 10:05:48.486664 4934 scope.go:117] "RemoveContainer" containerID="8d49c96ce7cdcce2add8644c0cd7f83e5d8576e97ea86ade9200b76b053d81e2" Oct 02 10:05:48 crc kubenswrapper[4934]: I1002 10:05:48.491390 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b5fff88f5-9wtdr"] Oct 02 10:05:48 crc kubenswrapper[4934]: I1002 10:05:48.497828 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b5fff88f5-9wtdr"] Oct 02 10:05:48 crc kubenswrapper[4934]: I1002 10:05:48.574383 4934 scope.go:117] "RemoveContainer" containerID="8d2579d409297d2b8abaf65ea5bb98ea1c241b6eda4c7c4f4d24d9d481f1a07c" Oct 02 10:05:48 crc kubenswrapper[4934]: E1002 10:05:48.574983 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d2579d409297d2b8abaf65ea5bb98ea1c241b6eda4c7c4f4d24d9d481f1a07c\": container with ID starting with 8d2579d409297d2b8abaf65ea5bb98ea1c241b6eda4c7c4f4d24d9d481f1a07c not found: ID does not exist" containerID="8d2579d409297d2b8abaf65ea5bb98ea1c241b6eda4c7c4f4d24d9d481f1a07c" Oct 02 10:05:48 crc kubenswrapper[4934]: I1002 10:05:48.575022 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d2579d409297d2b8abaf65ea5bb98ea1c241b6eda4c7c4f4d24d9d481f1a07c"} err="failed to get container status \"8d2579d409297d2b8abaf65ea5bb98ea1c241b6eda4c7c4f4d24d9d481f1a07c\": rpc error: code = NotFound desc = could not find container \"8d2579d409297d2b8abaf65ea5bb98ea1c241b6eda4c7c4f4d24d9d481f1a07c\": container with ID starting with 8d2579d409297d2b8abaf65ea5bb98ea1c241b6eda4c7c4f4d24d9d481f1a07c not found: ID does not exist" Oct 02 10:05:48 crc kubenswrapper[4934]: I1002 10:05:48.575074 4934 scope.go:117] "RemoveContainer" containerID="8d49c96ce7cdcce2add8644c0cd7f83e5d8576e97ea86ade9200b76b053d81e2" Oct 02 10:05:48 crc kubenswrapper[4934]: E1002 10:05:48.575422 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d49c96ce7cdcce2add8644c0cd7f83e5d8576e97ea86ade9200b76b053d81e2\": container with ID starting with 8d49c96ce7cdcce2add8644c0cd7f83e5d8576e97ea86ade9200b76b053d81e2 not found: ID does not exist" containerID="8d49c96ce7cdcce2add8644c0cd7f83e5d8576e97ea86ade9200b76b053d81e2" Oct 02 10:05:48 crc kubenswrapper[4934]: I1002 10:05:48.575446 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d49c96ce7cdcce2add8644c0cd7f83e5d8576e97ea86ade9200b76b053d81e2"} err="failed to get container status \"8d49c96ce7cdcce2add8644c0cd7f83e5d8576e97ea86ade9200b76b053d81e2\": rpc error: code = NotFound desc = could not find container \"8d49c96ce7cdcce2add8644c0cd7f83e5d8576e97ea86ade9200b76b053d81e2\": container with ID starting with 8d49c96ce7cdcce2add8644c0cd7f83e5d8576e97ea86ade9200b76b053d81e2 not found: ID does not exist" Oct 02 10:05:48 crc kubenswrapper[4934]: I1002 10:05:48.929943 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1ed141e-7d82-47c6-8d17-c1a1e1d601a7" path="/var/lib/kubelet/pods/a1ed141e-7d82-47c6-8d17-c1a1e1d601a7/volumes" Oct 02 10:05:48 crc kubenswrapper[4934]: I1002 10:05:48.931019 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8d5c788-c201-4daa-9b3c-3ea12cf696ff" path="/var/lib/kubelet/pods/d8d5c788-c201-4daa-9b3c-3ea12cf696ff/volumes" Oct 02 10:05:49 crc kubenswrapper[4934]: I1002 10:05:49.393987 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" event={"ID":"b3c7b1c5-61ea-4573-8709-252ea669dd2b","Type":"ContainerStarted","Data":"dc2c8e84c32ccaf851ee25f3bb96cdb382f84e514626c9fa94a7396f764dadd5"} Oct 02 10:05:49 crc kubenswrapper[4934]: I1002 10:05:49.394405 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" Oct 02 10:05:49 crc kubenswrapper[4934]: I1002 10:05:49.405519 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-769b765785-hsdtk" event={"ID":"0302150f-a71a-45f0-a6b4-024664b21e51","Type":"ContainerStarted","Data":"c021b99af4ef55634dacbbd18d21bee250069171cab0f36d2e5639562b6b4fa5"} Oct 02 10:05:49 crc kubenswrapper[4934]: I1002 10:05:49.405650 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-769b765785-hsdtk" Oct 02 10:05:49 crc kubenswrapper[4934]: I1002 10:05:49.411895 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"05c8faa0-6986-415d-abc3-4d35291375f2","Type":"ContainerStarted","Data":"6ab70322bf5ef2e1d2a4fd23bb06ac7d14bc671fa582c8397ceb83e14554def5"} Oct 02 10:05:49 crc kubenswrapper[4934]: I1002 10:05:49.421639 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" podStartSLOduration=3.421618493 podStartE2EDuration="3.421618493s" podCreationTimestamp="2025-10-02 10:05:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:05:49.420259454 +0000 UTC m=+1021.172900986" watchObservedRunningTime="2025-10-02 10:05:49.421618493 +0000 UTC m=+1021.174260015" Oct 02 10:05:49 crc kubenswrapper[4934]: I1002 10:05:49.449420 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-769b765785-hsdtk" podStartSLOduration=3.449398375 podStartE2EDuration="3.449398375s" podCreationTimestamp="2025-10-02 10:05:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:05:49.443006813 +0000 UTC m=+1021.195648335" watchObservedRunningTime="2025-10-02 10:05:49.449398375 +0000 UTC m=+1021.202039897" Oct 02 10:05:50 crc kubenswrapper[4934]: I1002 10:05:50.420589 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"05c8faa0-6986-415d-abc3-4d35291375f2","Type":"ContainerStarted","Data":"c0691673552ec68e68f4be61b4e9826705cb137abb4ac7f57faa0ea2d303510d"} Oct 02 10:05:50 crc kubenswrapper[4934]: I1002 10:05:50.439554 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.145546494 podStartE2EDuration="4.439537107s" podCreationTimestamp="2025-10-02 10:05:46 +0000 UTC" firstStartedPulling="2025-10-02 10:05:47.84587908 +0000 UTC m=+1019.598520622" lastFinishedPulling="2025-10-02 10:05:49.139869713 +0000 UTC m=+1020.892511235" observedRunningTime="2025-10-02 10:05:50.436902542 +0000 UTC m=+1022.189544064" watchObservedRunningTime="2025-10-02 10:05:50.439537107 +0000 UTC m=+1022.192178629" Oct 02 10:05:50 crc kubenswrapper[4934]: I1002 10:05:50.565893 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 02 10:05:50 crc kubenswrapper[4934]: I1002 10:05:50.566298 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 02 10:05:50 crc kubenswrapper[4934]: I1002 10:05:50.998789 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:50 crc kubenswrapper[4934]: I1002 10:05:50.998885 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:51 crc kubenswrapper[4934]: I1002 10:05:51.062072 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:51 crc kubenswrapper[4934]: I1002 10:05:51.430180 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 02 10:05:51 crc kubenswrapper[4934]: I1002 10:05:51.514376 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 02 10:05:52 crc kubenswrapper[4934]: I1002 10:05:52.676415 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 02 10:05:52 crc kubenswrapper[4934]: I1002 10:05:52.732814 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 02 10:05:53 crc kubenswrapper[4934]: I1002 10:05:53.359410 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 02 10:05:53 crc kubenswrapper[4934]: I1002 10:05:53.422986 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bd4b77fc7-w8tlj"] Oct 02 10:05:53 crc kubenswrapper[4934]: I1002 10:05:53.423250 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" podUID="b3c7b1c5-61ea-4573-8709-252ea669dd2b" containerName="dnsmasq-dns" containerID="cri-o://dc2c8e84c32ccaf851ee25f3bb96cdb382f84e514626c9fa94a7396f764dadd5" gracePeriod=10 Oct 02 10:05:53 crc kubenswrapper[4934]: I1002 10:05:53.433751 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" Oct 02 10:05:53 crc kubenswrapper[4934]: I1002 10:05:53.494243 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-556b48f7c7-8rms5"] Oct 02 10:05:53 crc kubenswrapper[4934]: E1002 10:05:53.495268 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8d5c788-c201-4daa-9b3c-3ea12cf696ff" containerName="init" Oct 02 10:05:53 crc kubenswrapper[4934]: I1002 10:05:53.495286 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8d5c788-c201-4daa-9b3c-3ea12cf696ff" containerName="init" Oct 02 10:05:53 crc kubenswrapper[4934]: E1002 10:05:53.495299 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8d5c788-c201-4daa-9b3c-3ea12cf696ff" containerName="dnsmasq-dns" Oct 02 10:05:53 crc kubenswrapper[4934]: I1002 10:05:53.495328 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8d5c788-c201-4daa-9b3c-3ea12cf696ff" containerName="dnsmasq-dns" Oct 02 10:05:53 crc kubenswrapper[4934]: I1002 10:05:53.498396 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8d5c788-c201-4daa-9b3c-3ea12cf696ff" containerName="dnsmasq-dns" Oct 02 10:05:53 crc kubenswrapper[4934]: I1002 10:05:53.499725 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" Oct 02 10:05:53 crc kubenswrapper[4934]: I1002 10:05:53.516652 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-556b48f7c7-8rms5"] Oct 02 10:05:53 crc kubenswrapper[4934]: I1002 10:05:53.638998 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/19ef1848-83f1-4a21-903c-ea7d36522dd2-ovsdbserver-sb\") pod \"dnsmasq-dns-556b48f7c7-8rms5\" (UID: \"19ef1848-83f1-4a21-903c-ea7d36522dd2\") " pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" Oct 02 10:05:53 crc kubenswrapper[4934]: I1002 10:05:53.639373 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19ef1848-83f1-4a21-903c-ea7d36522dd2-config\") pod \"dnsmasq-dns-556b48f7c7-8rms5\" (UID: \"19ef1848-83f1-4a21-903c-ea7d36522dd2\") " pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" Oct 02 10:05:53 crc kubenswrapper[4934]: I1002 10:05:53.639402 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvzm6\" (UniqueName: \"kubernetes.io/projected/19ef1848-83f1-4a21-903c-ea7d36522dd2-kube-api-access-xvzm6\") pod \"dnsmasq-dns-556b48f7c7-8rms5\" (UID: \"19ef1848-83f1-4a21-903c-ea7d36522dd2\") " pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" Oct 02 10:05:53 crc kubenswrapper[4934]: I1002 10:05:53.639456 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19ef1848-83f1-4a21-903c-ea7d36522dd2-ovsdbserver-nb\") pod \"dnsmasq-dns-556b48f7c7-8rms5\" (UID: \"19ef1848-83f1-4a21-903c-ea7d36522dd2\") " pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" Oct 02 10:05:53 crc kubenswrapper[4934]: I1002 10:05:53.639612 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19ef1848-83f1-4a21-903c-ea7d36522dd2-dns-svc\") pod \"dnsmasq-dns-556b48f7c7-8rms5\" (UID: \"19ef1848-83f1-4a21-903c-ea7d36522dd2\") " pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" Oct 02 10:05:53 crc kubenswrapper[4934]: I1002 10:05:53.740832 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19ef1848-83f1-4a21-903c-ea7d36522dd2-config\") pod \"dnsmasq-dns-556b48f7c7-8rms5\" (UID: \"19ef1848-83f1-4a21-903c-ea7d36522dd2\") " pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" Oct 02 10:05:53 crc kubenswrapper[4934]: I1002 10:05:53.740896 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvzm6\" (UniqueName: \"kubernetes.io/projected/19ef1848-83f1-4a21-903c-ea7d36522dd2-kube-api-access-xvzm6\") pod \"dnsmasq-dns-556b48f7c7-8rms5\" (UID: \"19ef1848-83f1-4a21-903c-ea7d36522dd2\") " pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" Oct 02 10:05:53 crc kubenswrapper[4934]: I1002 10:05:53.740951 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19ef1848-83f1-4a21-903c-ea7d36522dd2-ovsdbserver-nb\") pod \"dnsmasq-dns-556b48f7c7-8rms5\" (UID: \"19ef1848-83f1-4a21-903c-ea7d36522dd2\") " pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" Oct 02 10:05:53 crc kubenswrapper[4934]: I1002 10:05:53.740987 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19ef1848-83f1-4a21-903c-ea7d36522dd2-dns-svc\") pod \"dnsmasq-dns-556b48f7c7-8rms5\" (UID: \"19ef1848-83f1-4a21-903c-ea7d36522dd2\") " pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" Oct 02 10:05:53 crc kubenswrapper[4934]: I1002 10:05:53.741062 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/19ef1848-83f1-4a21-903c-ea7d36522dd2-ovsdbserver-sb\") pod \"dnsmasq-dns-556b48f7c7-8rms5\" (UID: \"19ef1848-83f1-4a21-903c-ea7d36522dd2\") " pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" Oct 02 10:05:53 crc kubenswrapper[4934]: I1002 10:05:53.741722 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19ef1848-83f1-4a21-903c-ea7d36522dd2-config\") pod \"dnsmasq-dns-556b48f7c7-8rms5\" (UID: \"19ef1848-83f1-4a21-903c-ea7d36522dd2\") " pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" Oct 02 10:05:53 crc kubenswrapper[4934]: I1002 10:05:53.741763 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19ef1848-83f1-4a21-903c-ea7d36522dd2-ovsdbserver-nb\") pod \"dnsmasq-dns-556b48f7c7-8rms5\" (UID: \"19ef1848-83f1-4a21-903c-ea7d36522dd2\") " pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" Oct 02 10:05:53 crc kubenswrapper[4934]: I1002 10:05:53.741915 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/19ef1848-83f1-4a21-903c-ea7d36522dd2-ovsdbserver-sb\") pod \"dnsmasq-dns-556b48f7c7-8rms5\" (UID: \"19ef1848-83f1-4a21-903c-ea7d36522dd2\") " pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" Oct 02 10:05:53 crc kubenswrapper[4934]: I1002 10:05:53.742053 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19ef1848-83f1-4a21-903c-ea7d36522dd2-dns-svc\") pod \"dnsmasq-dns-556b48f7c7-8rms5\" (UID: \"19ef1848-83f1-4a21-903c-ea7d36522dd2\") " pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" Oct 02 10:05:53 crc kubenswrapper[4934]: I1002 10:05:53.760162 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvzm6\" (UniqueName: \"kubernetes.io/projected/19ef1848-83f1-4a21-903c-ea7d36522dd2-kube-api-access-xvzm6\") pod \"dnsmasq-dns-556b48f7c7-8rms5\" (UID: \"19ef1848-83f1-4a21-903c-ea7d36522dd2\") " pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" Oct 02 10:05:53 crc kubenswrapper[4934]: I1002 10:05:53.866506 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" Oct 02 10:05:54 crc kubenswrapper[4934]: I1002 10:05:54.325689 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-556b48f7c7-8rms5"] Oct 02 10:05:54 crc kubenswrapper[4934]: I1002 10:05:54.455743 4934 generic.go:334] "Generic (PLEG): container finished" podID="b3c7b1c5-61ea-4573-8709-252ea669dd2b" containerID="dc2c8e84c32ccaf851ee25f3bb96cdb382f84e514626c9fa94a7396f764dadd5" exitCode=0 Oct 02 10:05:54 crc kubenswrapper[4934]: I1002 10:05:54.455803 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" event={"ID":"b3c7b1c5-61ea-4573-8709-252ea669dd2b","Type":"ContainerDied","Data":"dc2c8e84c32ccaf851ee25f3bb96cdb382f84e514626c9fa94a7396f764dadd5"} Oct 02 10:05:54 crc kubenswrapper[4934]: I1002 10:05:54.458424 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" event={"ID":"19ef1848-83f1-4a21-903c-ea7d36522dd2","Type":"ContainerStarted","Data":"1661fc9453c0b8d374163b8cdc07bc1248e6c719538fc4e79f19967449e38ebf"} Oct 02 10:05:54 crc kubenswrapper[4934]: I1002 10:05:54.593748 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 02 10:05:54 crc kubenswrapper[4934]: I1002 10:05:54.605877 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 02 10:05:54 crc kubenswrapper[4934]: I1002 10:05:54.607115 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 02 10:05:54 crc kubenswrapper[4934]: I1002 10:05:54.608784 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 02 10:05:54 crc kubenswrapper[4934]: I1002 10:05:54.609095 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 02 10:05:54 crc kubenswrapper[4934]: I1002 10:05:54.609318 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 02 10:05:54 crc kubenswrapper[4934]: I1002 10:05:54.610769 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-qsvk5" Oct 02 10:05:54 crc kubenswrapper[4934]: I1002 10:05:54.755664 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/8cebd152-5aaf-4543-be11-7ba327cfdf0e-cache\") pod \"swift-storage-0\" (UID: \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\") " pod="openstack/swift-storage-0" Oct 02 10:05:54 crc kubenswrapper[4934]: I1002 10:05:54.755772 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8cebd152-5aaf-4543-be11-7ba327cfdf0e-etc-swift\") pod \"swift-storage-0\" (UID: \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\") " pod="openstack/swift-storage-0" Oct 02 10:05:54 crc kubenswrapper[4934]: I1002 10:05:54.755808 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/8cebd152-5aaf-4543-be11-7ba327cfdf0e-lock\") pod \"swift-storage-0\" (UID: \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\") " pod="openstack/swift-storage-0" Oct 02 10:05:54 crc kubenswrapper[4934]: I1002 10:05:54.755837 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zt2qp\" (UniqueName: \"kubernetes.io/projected/8cebd152-5aaf-4543-be11-7ba327cfdf0e-kube-api-access-zt2qp\") pod \"swift-storage-0\" (UID: \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\") " pod="openstack/swift-storage-0" Oct 02 10:05:54 crc kubenswrapper[4934]: I1002 10:05:54.755878 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\") " pod="openstack/swift-storage-0" Oct 02 10:05:54 crc kubenswrapper[4934]: I1002 10:05:54.857503 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/8cebd152-5aaf-4543-be11-7ba327cfdf0e-cache\") pod \"swift-storage-0\" (UID: \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\") " pod="openstack/swift-storage-0" Oct 02 10:05:54 crc kubenswrapper[4934]: I1002 10:05:54.857658 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8cebd152-5aaf-4543-be11-7ba327cfdf0e-etc-swift\") pod \"swift-storage-0\" (UID: \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\") " pod="openstack/swift-storage-0" Oct 02 10:05:54 crc kubenswrapper[4934]: I1002 10:05:54.857709 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/8cebd152-5aaf-4543-be11-7ba327cfdf0e-lock\") pod \"swift-storage-0\" (UID: \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\") " pod="openstack/swift-storage-0" Oct 02 10:05:54 crc kubenswrapper[4934]: I1002 10:05:54.857752 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zt2qp\" (UniqueName: \"kubernetes.io/projected/8cebd152-5aaf-4543-be11-7ba327cfdf0e-kube-api-access-zt2qp\") pod \"swift-storage-0\" (UID: \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\") " pod="openstack/swift-storage-0" Oct 02 10:05:54 crc kubenswrapper[4934]: I1002 10:05:54.857789 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\") " pod="openstack/swift-storage-0" Oct 02 10:05:54 crc kubenswrapper[4934]: E1002 10:05:54.857894 4934 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 10:05:54 crc kubenswrapper[4934]: E1002 10:05:54.857932 4934 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 10:05:54 crc kubenswrapper[4934]: E1002 10:05:54.857997 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8cebd152-5aaf-4543-be11-7ba327cfdf0e-etc-swift podName:8cebd152-5aaf-4543-be11-7ba327cfdf0e nodeName:}" failed. No retries permitted until 2025-10-02 10:05:55.357977786 +0000 UTC m=+1027.110619318 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/8cebd152-5aaf-4543-be11-7ba327cfdf0e-etc-swift") pod "swift-storage-0" (UID: "8cebd152-5aaf-4543-be11-7ba327cfdf0e") : configmap "swift-ring-files" not found Oct 02 10:05:54 crc kubenswrapper[4934]: I1002 10:05:54.858118 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/8cebd152-5aaf-4543-be11-7ba327cfdf0e-cache\") pod \"swift-storage-0\" (UID: \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\") " pod="openstack/swift-storage-0" Oct 02 10:05:54 crc kubenswrapper[4934]: I1002 10:05:54.858169 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/swift-storage-0" Oct 02 10:05:54 crc kubenswrapper[4934]: I1002 10:05:54.858372 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/8cebd152-5aaf-4543-be11-7ba327cfdf0e-lock\") pod \"swift-storage-0\" (UID: \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\") " pod="openstack/swift-storage-0" Oct 02 10:05:54 crc kubenswrapper[4934]: I1002 10:05:54.884152 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\") " pod="openstack/swift-storage-0" Oct 02 10:05:54 crc kubenswrapper[4934]: I1002 10:05:54.889740 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zt2qp\" (UniqueName: \"kubernetes.io/projected/8cebd152-5aaf-4543-be11-7ba327cfdf0e-kube-api-access-zt2qp\") pod \"swift-storage-0\" (UID: \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\") " pod="openstack/swift-storage-0" Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.148514 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-n6t65"] Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.149899 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-n6t65" Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.151946 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.152265 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.162208 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-n6t65"] Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.163422 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.264412 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a1e933bf-e455-43c2-a996-b05a3bca3c96-swiftconf\") pod \"swift-ring-rebalance-n6t65\" (UID: \"a1e933bf-e455-43c2-a996-b05a3bca3c96\") " pod="openstack/swift-ring-rebalance-n6t65" Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.264518 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a1e933bf-e455-43c2-a996-b05a3bca3c96-dispersionconf\") pod \"swift-ring-rebalance-n6t65\" (UID: \"a1e933bf-e455-43c2-a996-b05a3bca3c96\") " pod="openstack/swift-ring-rebalance-n6t65" Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.264558 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1e933bf-e455-43c2-a996-b05a3bca3c96-combined-ca-bundle\") pod \"swift-ring-rebalance-n6t65\" (UID: \"a1e933bf-e455-43c2-a996-b05a3bca3c96\") " pod="openstack/swift-ring-rebalance-n6t65" Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.264629 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a1e933bf-e455-43c2-a996-b05a3bca3c96-etc-swift\") pod \"swift-ring-rebalance-n6t65\" (UID: \"a1e933bf-e455-43c2-a996-b05a3bca3c96\") " pod="openstack/swift-ring-rebalance-n6t65" Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.264723 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a1e933bf-e455-43c2-a996-b05a3bca3c96-ring-data-devices\") pod \"swift-ring-rebalance-n6t65\" (UID: \"a1e933bf-e455-43c2-a996-b05a3bca3c96\") " pod="openstack/swift-ring-rebalance-n6t65" Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.264783 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kchgx\" (UniqueName: \"kubernetes.io/projected/a1e933bf-e455-43c2-a996-b05a3bca3c96-kube-api-access-kchgx\") pod \"swift-ring-rebalance-n6t65\" (UID: \"a1e933bf-e455-43c2-a996-b05a3bca3c96\") " pod="openstack/swift-ring-rebalance-n6t65" Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.264829 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1e933bf-e455-43c2-a996-b05a3bca3c96-scripts\") pod \"swift-ring-rebalance-n6t65\" (UID: \"a1e933bf-e455-43c2-a996-b05a3bca3c96\") " pod="openstack/swift-ring-rebalance-n6t65" Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.366788 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a1e933bf-e455-43c2-a996-b05a3bca3c96-swiftconf\") pod \"swift-ring-rebalance-n6t65\" (UID: \"a1e933bf-e455-43c2-a996-b05a3bca3c96\") " pod="openstack/swift-ring-rebalance-n6t65" Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.366906 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8cebd152-5aaf-4543-be11-7ba327cfdf0e-etc-swift\") pod \"swift-storage-0\" (UID: \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\") " pod="openstack/swift-storage-0" Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.366964 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a1e933bf-e455-43c2-a996-b05a3bca3c96-dispersionconf\") pod \"swift-ring-rebalance-n6t65\" (UID: \"a1e933bf-e455-43c2-a996-b05a3bca3c96\") " pod="openstack/swift-ring-rebalance-n6t65" Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.367005 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1e933bf-e455-43c2-a996-b05a3bca3c96-combined-ca-bundle\") pod \"swift-ring-rebalance-n6t65\" (UID: \"a1e933bf-e455-43c2-a996-b05a3bca3c96\") " pod="openstack/swift-ring-rebalance-n6t65" Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.367051 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a1e933bf-e455-43c2-a996-b05a3bca3c96-etc-swift\") pod \"swift-ring-rebalance-n6t65\" (UID: \"a1e933bf-e455-43c2-a996-b05a3bca3c96\") " pod="openstack/swift-ring-rebalance-n6t65" Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.367096 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a1e933bf-e455-43c2-a996-b05a3bca3c96-ring-data-devices\") pod \"swift-ring-rebalance-n6t65\" (UID: \"a1e933bf-e455-43c2-a996-b05a3bca3c96\") " pod="openstack/swift-ring-rebalance-n6t65" Oct 02 10:05:55 crc kubenswrapper[4934]: E1002 10:05:55.367089 4934 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.367137 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kchgx\" (UniqueName: \"kubernetes.io/projected/a1e933bf-e455-43c2-a996-b05a3bca3c96-kube-api-access-kchgx\") pod \"swift-ring-rebalance-n6t65\" (UID: \"a1e933bf-e455-43c2-a996-b05a3bca3c96\") " pod="openstack/swift-ring-rebalance-n6t65" Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.367174 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1e933bf-e455-43c2-a996-b05a3bca3c96-scripts\") pod \"swift-ring-rebalance-n6t65\" (UID: \"a1e933bf-e455-43c2-a996-b05a3bca3c96\") " pod="openstack/swift-ring-rebalance-n6t65" Oct 02 10:05:55 crc kubenswrapper[4934]: E1002 10:05:55.367143 4934 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 10:05:55 crc kubenswrapper[4934]: E1002 10:05:55.367349 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8cebd152-5aaf-4543-be11-7ba327cfdf0e-etc-swift podName:8cebd152-5aaf-4543-be11-7ba327cfdf0e nodeName:}" failed. No retries permitted until 2025-10-02 10:05:56.367326844 +0000 UTC m=+1028.119968366 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/8cebd152-5aaf-4543-be11-7ba327cfdf0e-etc-swift") pod "swift-storage-0" (UID: "8cebd152-5aaf-4543-be11-7ba327cfdf0e") : configmap "swift-ring-files" not found Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.368145 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a1e933bf-e455-43c2-a996-b05a3bca3c96-etc-swift\") pod \"swift-ring-rebalance-n6t65\" (UID: \"a1e933bf-e455-43c2-a996-b05a3bca3c96\") " pod="openstack/swift-ring-rebalance-n6t65" Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.368182 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1e933bf-e455-43c2-a996-b05a3bca3c96-scripts\") pod \"swift-ring-rebalance-n6t65\" (UID: \"a1e933bf-e455-43c2-a996-b05a3bca3c96\") " pod="openstack/swift-ring-rebalance-n6t65" Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.368263 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a1e933bf-e455-43c2-a996-b05a3bca3c96-ring-data-devices\") pod \"swift-ring-rebalance-n6t65\" (UID: \"a1e933bf-e455-43c2-a996-b05a3bca3c96\") " pod="openstack/swift-ring-rebalance-n6t65" Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.370375 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a1e933bf-e455-43c2-a996-b05a3bca3c96-swiftconf\") pod \"swift-ring-rebalance-n6t65\" (UID: \"a1e933bf-e455-43c2-a996-b05a3bca3c96\") " pod="openstack/swift-ring-rebalance-n6t65" Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.370895 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1e933bf-e455-43c2-a996-b05a3bca3c96-combined-ca-bundle\") pod \"swift-ring-rebalance-n6t65\" (UID: \"a1e933bf-e455-43c2-a996-b05a3bca3c96\") " pod="openstack/swift-ring-rebalance-n6t65" Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.372469 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a1e933bf-e455-43c2-a996-b05a3bca3c96-dispersionconf\") pod \"swift-ring-rebalance-n6t65\" (UID: \"a1e933bf-e455-43c2-a996-b05a3bca3c96\") " pod="openstack/swift-ring-rebalance-n6t65" Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.397021 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kchgx\" (UniqueName: \"kubernetes.io/projected/a1e933bf-e455-43c2-a996-b05a3bca3c96-kube-api-access-kchgx\") pod \"swift-ring-rebalance-n6t65\" (UID: \"a1e933bf-e455-43c2-a996-b05a3bca3c96\") " pod="openstack/swift-ring-rebalance-n6t65" Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.474351 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-n6t65" Oct 02 10:05:55 crc kubenswrapper[4934]: I1002 10:05:55.918940 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-n6t65"] Oct 02 10:05:55 crc kubenswrapper[4934]: W1002 10:05:55.921999 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1e933bf_e455_43c2_a996_b05a3bca3c96.slice/crio-618e539486eeeb7909a3130d7aef5c1e7f43783aabc80d44b99d43f4e73b6cf2 WatchSource:0}: Error finding container 618e539486eeeb7909a3130d7aef5c1e7f43783aabc80d44b99d43f4e73b6cf2: Status 404 returned error can't find the container with id 618e539486eeeb7909a3130d7aef5c1e7f43783aabc80d44b99d43f4e73b6cf2 Oct 02 10:05:56 crc kubenswrapper[4934]: I1002 10:05:56.384812 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8cebd152-5aaf-4543-be11-7ba327cfdf0e-etc-swift\") pod \"swift-storage-0\" (UID: \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\") " pod="openstack/swift-storage-0" Oct 02 10:05:56 crc kubenswrapper[4934]: E1002 10:05:56.385221 4934 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 10:05:56 crc kubenswrapper[4934]: E1002 10:05:56.385257 4934 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 10:05:56 crc kubenswrapper[4934]: E1002 10:05:56.385334 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8cebd152-5aaf-4543-be11-7ba327cfdf0e-etc-swift podName:8cebd152-5aaf-4543-be11-7ba327cfdf0e nodeName:}" failed. No retries permitted until 2025-10-02 10:05:58.385310359 +0000 UTC m=+1030.137951921 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/8cebd152-5aaf-4543-be11-7ba327cfdf0e-etc-swift") pod "swift-storage-0" (UID: "8cebd152-5aaf-4543-be11-7ba327cfdf0e") : configmap "swift-ring-files" not found Oct 02 10:05:56 crc kubenswrapper[4934]: I1002 10:05:56.496919 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-n6t65" event={"ID":"a1e933bf-e455-43c2-a996-b05a3bca3c96","Type":"ContainerStarted","Data":"618e539486eeeb7909a3130d7aef5c1e7f43783aabc80d44b99d43f4e73b6cf2"} Oct 02 10:05:56 crc kubenswrapper[4934]: I1002 10:05:56.748255 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-cwxjc"] Oct 02 10:05:56 crc kubenswrapper[4934]: I1002 10:05:56.749509 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-cwxjc" Oct 02 10:05:56 crc kubenswrapper[4934]: I1002 10:05:56.766204 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-cwxjc"] Oct 02 10:05:56 crc kubenswrapper[4934]: I1002 10:05:56.901608 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njn5q\" (UniqueName: \"kubernetes.io/projected/46721f65-4fa6-4059-aeb2-c9e62ed22cb1-kube-api-access-njn5q\") pod \"glance-db-create-cwxjc\" (UID: \"46721f65-4fa6-4059-aeb2-c9e62ed22cb1\") " pod="openstack/glance-db-create-cwxjc" Oct 02 10:05:57 crc kubenswrapper[4934]: I1002 10:05:57.003334 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njn5q\" (UniqueName: \"kubernetes.io/projected/46721f65-4fa6-4059-aeb2-c9e62ed22cb1-kube-api-access-njn5q\") pod \"glance-db-create-cwxjc\" (UID: \"46721f65-4fa6-4059-aeb2-c9e62ed22cb1\") " pod="openstack/glance-db-create-cwxjc" Oct 02 10:05:57 crc kubenswrapper[4934]: I1002 10:05:57.028545 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njn5q\" (UniqueName: \"kubernetes.io/projected/46721f65-4fa6-4059-aeb2-c9e62ed22cb1-kube-api-access-njn5q\") pod \"glance-db-create-cwxjc\" (UID: \"46721f65-4fa6-4059-aeb2-c9e62ed22cb1\") " pod="openstack/glance-db-create-cwxjc" Oct 02 10:05:57 crc kubenswrapper[4934]: I1002 10:05:57.053761 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-769b765785-hsdtk" Oct 02 10:05:57 crc kubenswrapper[4934]: I1002 10:05:57.069872 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-cwxjc" Oct 02 10:05:57 crc kubenswrapper[4934]: I1002 10:05:57.300714 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" podUID="b3c7b1c5-61ea-4573-8709-252ea669dd2b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.115:5353: connect: connection refused" Oct 02 10:05:57 crc kubenswrapper[4934]: I1002 10:05:57.499509 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-cwxjc"] Oct 02 10:05:58 crc kubenswrapper[4934]: I1002 10:05:58.426442 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8cebd152-5aaf-4543-be11-7ba327cfdf0e-etc-swift\") pod \"swift-storage-0\" (UID: \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\") " pod="openstack/swift-storage-0" Oct 02 10:05:58 crc kubenswrapper[4934]: E1002 10:05:58.426663 4934 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 10:05:58 crc kubenswrapper[4934]: E1002 10:05:58.428269 4934 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 10:05:58 crc kubenswrapper[4934]: E1002 10:05:58.428451 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8cebd152-5aaf-4543-be11-7ba327cfdf0e-etc-swift podName:8cebd152-5aaf-4543-be11-7ba327cfdf0e nodeName:}" failed. No retries permitted until 2025-10-02 10:06:02.428426975 +0000 UTC m=+1034.181068517 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/8cebd152-5aaf-4543-be11-7ba327cfdf0e-etc-swift") pod "swift-storage-0" (UID: "8cebd152-5aaf-4543-be11-7ba327cfdf0e") : configmap "swift-ring-files" not found Oct 02 10:05:58 crc kubenswrapper[4934]: I1002 10:05:58.516045 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-cwxjc" event={"ID":"46721f65-4fa6-4059-aeb2-c9e62ed22cb1","Type":"ContainerStarted","Data":"a68289a98ecbf081d7bfa8a6a2c70863a26f27bd993f0f037307e6868bbb188e"} Oct 02 10:05:59 crc kubenswrapper[4934]: I1002 10:05:59.984471 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" Oct 02 10:06:00 crc kubenswrapper[4934]: I1002 10:06:00.158972 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3c7b1c5-61ea-4573-8709-252ea669dd2b-ovsdbserver-sb\") pod \"b3c7b1c5-61ea-4573-8709-252ea669dd2b\" (UID: \"b3c7b1c5-61ea-4573-8709-252ea669dd2b\") " Oct 02 10:06:00 crc kubenswrapper[4934]: I1002 10:06:00.159063 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3c7b1c5-61ea-4573-8709-252ea669dd2b-dns-svc\") pod \"b3c7b1c5-61ea-4573-8709-252ea669dd2b\" (UID: \"b3c7b1c5-61ea-4573-8709-252ea669dd2b\") " Oct 02 10:06:00 crc kubenswrapper[4934]: I1002 10:06:00.159162 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldqdc\" (UniqueName: \"kubernetes.io/projected/b3c7b1c5-61ea-4573-8709-252ea669dd2b-kube-api-access-ldqdc\") pod \"b3c7b1c5-61ea-4573-8709-252ea669dd2b\" (UID: \"b3c7b1c5-61ea-4573-8709-252ea669dd2b\") " Oct 02 10:06:00 crc kubenswrapper[4934]: I1002 10:06:00.159200 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3c7b1c5-61ea-4573-8709-252ea669dd2b-ovsdbserver-nb\") pod \"b3c7b1c5-61ea-4573-8709-252ea669dd2b\" (UID: \"b3c7b1c5-61ea-4573-8709-252ea669dd2b\") " Oct 02 10:06:00 crc kubenswrapper[4934]: I1002 10:06:00.159237 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3c7b1c5-61ea-4573-8709-252ea669dd2b-config\") pod \"b3c7b1c5-61ea-4573-8709-252ea669dd2b\" (UID: \"b3c7b1c5-61ea-4573-8709-252ea669dd2b\") " Oct 02 10:06:00 crc kubenswrapper[4934]: I1002 10:06:00.164794 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3c7b1c5-61ea-4573-8709-252ea669dd2b-kube-api-access-ldqdc" (OuterVolumeSpecName: "kube-api-access-ldqdc") pod "b3c7b1c5-61ea-4573-8709-252ea669dd2b" (UID: "b3c7b1c5-61ea-4573-8709-252ea669dd2b"). InnerVolumeSpecName "kube-api-access-ldqdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:00 crc kubenswrapper[4934]: I1002 10:06:00.197699 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3c7b1c5-61ea-4573-8709-252ea669dd2b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b3c7b1c5-61ea-4573-8709-252ea669dd2b" (UID: "b3c7b1c5-61ea-4573-8709-252ea669dd2b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:06:00 crc kubenswrapper[4934]: I1002 10:06:00.202094 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3c7b1c5-61ea-4573-8709-252ea669dd2b-config" (OuterVolumeSpecName: "config") pod "b3c7b1c5-61ea-4573-8709-252ea669dd2b" (UID: "b3c7b1c5-61ea-4573-8709-252ea669dd2b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:06:00 crc kubenswrapper[4934]: I1002 10:06:00.204890 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3c7b1c5-61ea-4573-8709-252ea669dd2b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b3c7b1c5-61ea-4573-8709-252ea669dd2b" (UID: "b3c7b1c5-61ea-4573-8709-252ea669dd2b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:06:00 crc kubenswrapper[4934]: I1002 10:06:00.214309 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3c7b1c5-61ea-4573-8709-252ea669dd2b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b3c7b1c5-61ea-4573-8709-252ea669dd2b" (UID: "b3c7b1c5-61ea-4573-8709-252ea669dd2b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:06:00 crc kubenswrapper[4934]: I1002 10:06:00.261310 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3c7b1c5-61ea-4573-8709-252ea669dd2b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:00 crc kubenswrapper[4934]: I1002 10:06:00.261349 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3c7b1c5-61ea-4573-8709-252ea669dd2b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:00 crc kubenswrapper[4934]: I1002 10:06:00.261359 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldqdc\" (UniqueName: \"kubernetes.io/projected/b3c7b1c5-61ea-4573-8709-252ea669dd2b-kube-api-access-ldqdc\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:00 crc kubenswrapper[4934]: I1002 10:06:00.261370 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3c7b1c5-61ea-4573-8709-252ea669dd2b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:00 crc kubenswrapper[4934]: I1002 10:06:00.261379 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3c7b1c5-61ea-4573-8709-252ea669dd2b-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:00 crc kubenswrapper[4934]: I1002 10:06:00.535899 4934 generic.go:334] "Generic (PLEG): container finished" podID="46721f65-4fa6-4059-aeb2-c9e62ed22cb1" containerID="09ddc37d1aa16f33abd00be6bd097bcc1c45216a18e235121272e97806bb67e7" exitCode=0 Oct 02 10:06:00 crc kubenswrapper[4934]: I1002 10:06:00.536041 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-cwxjc" event={"ID":"46721f65-4fa6-4059-aeb2-c9e62ed22cb1","Type":"ContainerDied","Data":"09ddc37d1aa16f33abd00be6bd097bcc1c45216a18e235121272e97806bb67e7"} Oct 02 10:06:00 crc kubenswrapper[4934]: I1002 10:06:00.539032 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" Oct 02 10:06:00 crc kubenswrapper[4934]: I1002 10:06:00.539069 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bd4b77fc7-w8tlj" event={"ID":"b3c7b1c5-61ea-4573-8709-252ea669dd2b","Type":"ContainerDied","Data":"ee7cf74e3d31a22106b015efac21da25bde0e7f6d4ea31b2475764bfd094a780"} Oct 02 10:06:00 crc kubenswrapper[4934]: I1002 10:06:00.539122 4934 scope.go:117] "RemoveContainer" containerID="dc2c8e84c32ccaf851ee25f3bb96cdb382f84e514626c9fa94a7396f764dadd5" Oct 02 10:06:00 crc kubenswrapper[4934]: I1002 10:06:00.543262 4934 generic.go:334] "Generic (PLEG): container finished" podID="19ef1848-83f1-4a21-903c-ea7d36522dd2" containerID="076596e447df636fd2be21ec0842225ebdff1f702645ba2e8782db56398c415a" exitCode=0 Oct 02 10:06:00 crc kubenswrapper[4934]: I1002 10:06:00.543308 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" event={"ID":"19ef1848-83f1-4a21-903c-ea7d36522dd2","Type":"ContainerDied","Data":"076596e447df636fd2be21ec0842225ebdff1f702645ba2e8782db56398c415a"} Oct 02 10:06:00 crc kubenswrapper[4934]: I1002 10:06:00.591448 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bd4b77fc7-w8tlj"] Oct 02 10:06:00 crc kubenswrapper[4934]: I1002 10:06:00.598027 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bd4b77fc7-w8tlj"] Oct 02 10:06:00 crc kubenswrapper[4934]: I1002 10:06:00.924138 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3c7b1c5-61ea-4573-8709-252ea669dd2b" path="/var/lib/kubelet/pods/b3c7b1c5-61ea-4573-8709-252ea669dd2b/volumes" Oct 02 10:06:01 crc kubenswrapper[4934]: I1002 10:06:01.041292 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-mrnhj"] Oct 02 10:06:01 crc kubenswrapper[4934]: E1002 10:06:01.042611 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3c7b1c5-61ea-4573-8709-252ea669dd2b" containerName="init" Oct 02 10:06:01 crc kubenswrapper[4934]: I1002 10:06:01.042634 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3c7b1c5-61ea-4573-8709-252ea669dd2b" containerName="init" Oct 02 10:06:01 crc kubenswrapper[4934]: E1002 10:06:01.042646 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3c7b1c5-61ea-4573-8709-252ea669dd2b" containerName="dnsmasq-dns" Oct 02 10:06:01 crc kubenswrapper[4934]: I1002 10:06:01.042653 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3c7b1c5-61ea-4573-8709-252ea669dd2b" containerName="dnsmasq-dns" Oct 02 10:06:01 crc kubenswrapper[4934]: I1002 10:06:01.042899 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3c7b1c5-61ea-4573-8709-252ea669dd2b" containerName="dnsmasq-dns" Oct 02 10:06:01 crc kubenswrapper[4934]: I1002 10:06:01.043541 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-mrnhj" Oct 02 10:06:01 crc kubenswrapper[4934]: I1002 10:06:01.102341 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-mrnhj"] Oct 02 10:06:01 crc kubenswrapper[4934]: I1002 10:06:01.195192 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ts2ck\" (UniqueName: \"kubernetes.io/projected/4c705cde-0601-4f0e-97a3-590adbf0da4c-kube-api-access-ts2ck\") pod \"keystone-db-create-mrnhj\" (UID: \"4c705cde-0601-4f0e-97a3-590adbf0da4c\") " pod="openstack/keystone-db-create-mrnhj" Oct 02 10:06:01 crc kubenswrapper[4934]: I1002 10:06:01.296623 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ts2ck\" (UniqueName: \"kubernetes.io/projected/4c705cde-0601-4f0e-97a3-590adbf0da4c-kube-api-access-ts2ck\") pod \"keystone-db-create-mrnhj\" (UID: \"4c705cde-0601-4f0e-97a3-590adbf0da4c\") " pod="openstack/keystone-db-create-mrnhj" Oct 02 10:06:01 crc kubenswrapper[4934]: I1002 10:06:01.313768 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ts2ck\" (UniqueName: \"kubernetes.io/projected/4c705cde-0601-4f0e-97a3-590adbf0da4c-kube-api-access-ts2ck\") pod \"keystone-db-create-mrnhj\" (UID: \"4c705cde-0601-4f0e-97a3-590adbf0da4c\") " pod="openstack/keystone-db-create-mrnhj" Oct 02 10:06:01 crc kubenswrapper[4934]: I1002 10:06:01.332713 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-74sr7"] Oct 02 10:06:01 crc kubenswrapper[4934]: I1002 10:06:01.333659 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-74sr7" Oct 02 10:06:01 crc kubenswrapper[4934]: I1002 10:06:01.340254 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-74sr7"] Oct 02 10:06:01 crc kubenswrapper[4934]: I1002 10:06:01.411916 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-mrnhj" Oct 02 10:06:01 crc kubenswrapper[4934]: I1002 10:06:01.500523 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8dvc\" (UniqueName: \"kubernetes.io/projected/fe286dcd-593e-48ee-8140-aca585425610-kube-api-access-l8dvc\") pod \"placement-db-create-74sr7\" (UID: \"fe286dcd-593e-48ee-8140-aca585425610\") " pod="openstack/placement-db-create-74sr7" Oct 02 10:06:01 crc kubenswrapper[4934]: I1002 10:06:01.601993 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8dvc\" (UniqueName: \"kubernetes.io/projected/fe286dcd-593e-48ee-8140-aca585425610-kube-api-access-l8dvc\") pod \"placement-db-create-74sr7\" (UID: \"fe286dcd-593e-48ee-8140-aca585425610\") " pod="openstack/placement-db-create-74sr7" Oct 02 10:06:01 crc kubenswrapper[4934]: I1002 10:06:01.621671 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8dvc\" (UniqueName: \"kubernetes.io/projected/fe286dcd-593e-48ee-8140-aca585425610-kube-api-access-l8dvc\") pod \"placement-db-create-74sr7\" (UID: \"fe286dcd-593e-48ee-8140-aca585425610\") " pod="openstack/placement-db-create-74sr7" Oct 02 10:06:01 crc kubenswrapper[4934]: I1002 10:06:01.683744 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-74sr7" Oct 02 10:06:02 crc kubenswrapper[4934]: I1002 10:06:02.246140 4934 scope.go:117] "RemoveContainer" containerID="f29bffa051f3ee598a217da3e18a3447a85c5230a3130c9f68e351e0b0e00403" Oct 02 10:06:02 crc kubenswrapper[4934]: I1002 10:06:02.266934 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 02 10:06:02 crc kubenswrapper[4934]: I1002 10:06:02.522357 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8cebd152-5aaf-4543-be11-7ba327cfdf0e-etc-swift\") pod \"swift-storage-0\" (UID: \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\") " pod="openstack/swift-storage-0" Oct 02 10:06:02 crc kubenswrapper[4934]: E1002 10:06:02.522616 4934 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 10:06:02 crc kubenswrapper[4934]: E1002 10:06:02.522783 4934 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 10:06:02 crc kubenswrapper[4934]: E1002 10:06:02.522867 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8cebd152-5aaf-4543-be11-7ba327cfdf0e-etc-swift podName:8cebd152-5aaf-4543-be11-7ba327cfdf0e nodeName:}" failed. No retries permitted until 2025-10-02 10:06:10.522846419 +0000 UTC m=+1042.275487941 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/8cebd152-5aaf-4543-be11-7ba327cfdf0e-etc-swift") pod "swift-storage-0" (UID: "8cebd152-5aaf-4543-be11-7ba327cfdf0e") : configmap "swift-ring-files" not found Oct 02 10:06:02 crc kubenswrapper[4934]: I1002 10:06:02.564125 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-cwxjc" event={"ID":"46721f65-4fa6-4059-aeb2-c9e62ed22cb1","Type":"ContainerDied","Data":"a68289a98ecbf081d7bfa8a6a2c70863a26f27bd993f0f037307e6868bbb188e"} Oct 02 10:06:02 crc kubenswrapper[4934]: I1002 10:06:02.564162 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a68289a98ecbf081d7bfa8a6a2c70863a26f27bd993f0f037307e6868bbb188e" Oct 02 10:06:02 crc kubenswrapper[4934]: I1002 10:06:02.586032 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-cwxjc" Oct 02 10:06:02 crc kubenswrapper[4934]: I1002 10:06:02.624160 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njn5q\" (UniqueName: \"kubernetes.io/projected/46721f65-4fa6-4059-aeb2-c9e62ed22cb1-kube-api-access-njn5q\") pod \"46721f65-4fa6-4059-aeb2-c9e62ed22cb1\" (UID: \"46721f65-4fa6-4059-aeb2-c9e62ed22cb1\") " Oct 02 10:06:02 crc kubenswrapper[4934]: I1002 10:06:02.631808 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46721f65-4fa6-4059-aeb2-c9e62ed22cb1-kube-api-access-njn5q" (OuterVolumeSpecName: "kube-api-access-njn5q") pod "46721f65-4fa6-4059-aeb2-c9e62ed22cb1" (UID: "46721f65-4fa6-4059-aeb2-c9e62ed22cb1"). InnerVolumeSpecName "kube-api-access-njn5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:02 crc kubenswrapper[4934]: I1002 10:06:02.656844 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-74sr7"] Oct 02 10:06:02 crc kubenswrapper[4934]: W1002 10:06:02.657613 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe286dcd_593e_48ee_8140_aca585425610.slice/crio-68ff15951dead6d743a877a8282e925bbb498b161f67890c134953e985277ce0 WatchSource:0}: Error finding container 68ff15951dead6d743a877a8282e925bbb498b161f67890c134953e985277ce0: Status 404 returned error can't find the container with id 68ff15951dead6d743a877a8282e925bbb498b161f67890c134953e985277ce0 Oct 02 10:06:02 crc kubenswrapper[4934]: I1002 10:06:02.725825 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njn5q\" (UniqueName: \"kubernetes.io/projected/46721f65-4fa6-4059-aeb2-c9e62ed22cb1-kube-api-access-njn5q\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:02 crc kubenswrapper[4934]: I1002 10:06:02.793348 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-mrnhj"] Oct 02 10:06:02 crc kubenswrapper[4934]: W1002 10:06:02.794440 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c705cde_0601_4f0e_97a3_590adbf0da4c.slice/crio-c1cd08311f7a5688996c77f3dcf2e72e98b641246197b2aa3bc4d5f4d8916e77 WatchSource:0}: Error finding container c1cd08311f7a5688996c77f3dcf2e72e98b641246197b2aa3bc4d5f4d8916e77: Status 404 returned error can't find the container with id c1cd08311f7a5688996c77f3dcf2e72e98b641246197b2aa3bc4d5f4d8916e77 Oct 02 10:06:03 crc kubenswrapper[4934]: I1002 10:06:03.578032 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-n6t65" event={"ID":"a1e933bf-e455-43c2-a996-b05a3bca3c96","Type":"ContainerStarted","Data":"7826f59cb7f15b3e25a5cbac82bed398ef57ad483fbbed25d37b37218b482cf8"} Oct 02 10:06:03 crc kubenswrapper[4934]: I1002 10:06:03.580693 4934 generic.go:334] "Generic (PLEG): container finished" podID="fe286dcd-593e-48ee-8140-aca585425610" containerID="3c4a89f33ba516b7c1f812218bd4d35a373ceb64cc4479e97e705b984e4610db" exitCode=0 Oct 02 10:06:03 crc kubenswrapper[4934]: I1002 10:06:03.580821 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-74sr7" event={"ID":"fe286dcd-593e-48ee-8140-aca585425610","Type":"ContainerDied","Data":"3c4a89f33ba516b7c1f812218bd4d35a373ceb64cc4479e97e705b984e4610db"} Oct 02 10:06:03 crc kubenswrapper[4934]: I1002 10:06:03.580869 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-74sr7" event={"ID":"fe286dcd-593e-48ee-8140-aca585425610","Type":"ContainerStarted","Data":"68ff15951dead6d743a877a8282e925bbb498b161f67890c134953e985277ce0"} Oct 02 10:06:03 crc kubenswrapper[4934]: I1002 10:06:03.583053 4934 generic.go:334] "Generic (PLEG): container finished" podID="4c705cde-0601-4f0e-97a3-590adbf0da4c" containerID="b74f83fb8d92680f011c564f24d2339dcb490fefe27cf43645816a56cbb27d57" exitCode=0 Oct 02 10:06:03 crc kubenswrapper[4934]: I1002 10:06:03.583130 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-mrnhj" event={"ID":"4c705cde-0601-4f0e-97a3-590adbf0da4c","Type":"ContainerDied","Data":"b74f83fb8d92680f011c564f24d2339dcb490fefe27cf43645816a56cbb27d57"} Oct 02 10:06:03 crc kubenswrapper[4934]: I1002 10:06:03.583163 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-mrnhj" event={"ID":"4c705cde-0601-4f0e-97a3-590adbf0da4c","Type":"ContainerStarted","Data":"c1cd08311f7a5688996c77f3dcf2e72e98b641246197b2aa3bc4d5f4d8916e77"} Oct 02 10:06:03 crc kubenswrapper[4934]: I1002 10:06:03.585170 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-cwxjc" Oct 02 10:06:03 crc kubenswrapper[4934]: I1002 10:06:03.585518 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" event={"ID":"19ef1848-83f1-4a21-903c-ea7d36522dd2","Type":"ContainerStarted","Data":"5f23045960ea86318590a91a23756e660012b01bafb2ad070f7ad444363b5434"} Oct 02 10:06:03 crc kubenswrapper[4934]: I1002 10:06:03.585761 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" Oct 02 10:06:03 crc kubenswrapper[4934]: I1002 10:06:03.618390 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-n6t65" podStartSLOduration=2.194617747 podStartE2EDuration="8.618361735s" podCreationTimestamp="2025-10-02 10:05:55 +0000 UTC" firstStartedPulling="2025-10-02 10:05:55.924799003 +0000 UTC m=+1027.677440525" lastFinishedPulling="2025-10-02 10:06:02.348542981 +0000 UTC m=+1034.101184513" observedRunningTime="2025-10-02 10:06:03.606838866 +0000 UTC m=+1035.359480428" watchObservedRunningTime="2025-10-02 10:06:03.618361735 +0000 UTC m=+1035.371003277" Oct 02 10:06:03 crc kubenswrapper[4934]: I1002 10:06:03.636958 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" podStartSLOduration=10.636884493 podStartE2EDuration="10.636884493s" podCreationTimestamp="2025-10-02 10:05:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:06:03.632675333 +0000 UTC m=+1035.385316875" watchObservedRunningTime="2025-10-02 10:06:03.636884493 +0000 UTC m=+1035.389526025" Oct 02 10:06:04 crc kubenswrapper[4934]: I1002 10:06:04.931884 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-mrnhj" Oct 02 10:06:04 crc kubenswrapper[4934]: I1002 10:06:04.937738 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-74sr7" Oct 02 10:06:04 crc kubenswrapper[4934]: I1002 10:06:04.977162 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8dvc\" (UniqueName: \"kubernetes.io/projected/fe286dcd-593e-48ee-8140-aca585425610-kube-api-access-l8dvc\") pod \"fe286dcd-593e-48ee-8140-aca585425610\" (UID: \"fe286dcd-593e-48ee-8140-aca585425610\") " Oct 02 10:06:04 crc kubenswrapper[4934]: I1002 10:06:04.977241 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ts2ck\" (UniqueName: \"kubernetes.io/projected/4c705cde-0601-4f0e-97a3-590adbf0da4c-kube-api-access-ts2ck\") pod \"4c705cde-0601-4f0e-97a3-590adbf0da4c\" (UID: \"4c705cde-0601-4f0e-97a3-590adbf0da4c\") " Oct 02 10:06:04 crc kubenswrapper[4934]: I1002 10:06:04.982497 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe286dcd-593e-48ee-8140-aca585425610-kube-api-access-l8dvc" (OuterVolumeSpecName: "kube-api-access-l8dvc") pod "fe286dcd-593e-48ee-8140-aca585425610" (UID: "fe286dcd-593e-48ee-8140-aca585425610"). InnerVolumeSpecName "kube-api-access-l8dvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:04 crc kubenswrapper[4934]: I1002 10:06:04.982627 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c705cde-0601-4f0e-97a3-590adbf0da4c-kube-api-access-ts2ck" (OuterVolumeSpecName: "kube-api-access-ts2ck") pod "4c705cde-0601-4f0e-97a3-590adbf0da4c" (UID: "4c705cde-0601-4f0e-97a3-590adbf0da4c"). InnerVolumeSpecName "kube-api-access-ts2ck". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:05 crc kubenswrapper[4934]: I1002 10:06:05.081497 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8dvc\" (UniqueName: \"kubernetes.io/projected/fe286dcd-593e-48ee-8140-aca585425610-kube-api-access-l8dvc\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:05 crc kubenswrapper[4934]: I1002 10:06:05.081533 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ts2ck\" (UniqueName: \"kubernetes.io/projected/4c705cde-0601-4f0e-97a3-590adbf0da4c-kube-api-access-ts2ck\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:05 crc kubenswrapper[4934]: I1002 10:06:05.602855 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-74sr7" event={"ID":"fe286dcd-593e-48ee-8140-aca585425610","Type":"ContainerDied","Data":"68ff15951dead6d743a877a8282e925bbb498b161f67890c134953e985277ce0"} Oct 02 10:06:05 crc kubenswrapper[4934]: I1002 10:06:05.602913 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68ff15951dead6d743a877a8282e925bbb498b161f67890c134953e985277ce0" Oct 02 10:06:05 crc kubenswrapper[4934]: I1002 10:06:05.602994 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-74sr7" Oct 02 10:06:05 crc kubenswrapper[4934]: I1002 10:06:05.609617 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-mrnhj" event={"ID":"4c705cde-0601-4f0e-97a3-590adbf0da4c","Type":"ContainerDied","Data":"c1cd08311f7a5688996c77f3dcf2e72e98b641246197b2aa3bc4d5f4d8916e77"} Oct 02 10:06:05 crc kubenswrapper[4934]: I1002 10:06:05.609648 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1cd08311f7a5688996c77f3dcf2e72e98b641246197b2aa3bc4d5f4d8916e77" Oct 02 10:06:05 crc kubenswrapper[4934]: I1002 10:06:05.609709 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-mrnhj" Oct 02 10:06:06 crc kubenswrapper[4934]: I1002 10:06:06.621318 4934 generic.go:334] "Generic (PLEG): container finished" podID="97560469-c459-48f4-901e-54eaded32537" containerID="91a99821c085c0ca3aac996ae178fce6d284a5d09eab41f0c4748d65f0d4756d" exitCode=0 Oct 02 10:06:06 crc kubenswrapper[4934]: I1002 10:06:06.621399 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"97560469-c459-48f4-901e-54eaded32537","Type":"ContainerDied","Data":"91a99821c085c0ca3aac996ae178fce6d284a5d09eab41f0c4748d65f0d4756d"} Oct 02 10:06:07 crc kubenswrapper[4934]: I1002 10:06:07.632857 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"97560469-c459-48f4-901e-54eaded32537","Type":"ContainerStarted","Data":"2432f321bd3453c2b78b0dac285ee73ee7b8165fefef31515a2963451230f5a8"} Oct 02 10:06:07 crc kubenswrapper[4934]: I1002 10:06:07.634897 4934 generic.go:334] "Generic (PLEG): container finished" podID="d2b6e083-a523-48e2-8157-088036174e12" containerID="eaa7cefa7a244d7ebc3161825f4958eb21752c3953ea843f3cde9449733066de" exitCode=0 Oct 02 10:06:07 crc kubenswrapper[4934]: I1002 10:06:07.634934 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d2b6e083-a523-48e2-8157-088036174e12","Type":"ContainerDied","Data":"eaa7cefa7a244d7ebc3161825f4958eb21752c3953ea843f3cde9449733066de"} Oct 02 10:06:07 crc kubenswrapper[4934]: I1002 10:06:07.678006 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.298132517 podStartE2EDuration="51.677975886s" podCreationTimestamp="2025-10-02 10:05:16 +0000 UTC" firstStartedPulling="2025-10-02 10:05:19.083550787 +0000 UTC m=+990.836192309" lastFinishedPulling="2025-10-02 10:05:33.463394156 +0000 UTC m=+1005.216035678" observedRunningTime="2025-10-02 10:06:07.664656306 +0000 UTC m=+1039.417297898" watchObservedRunningTime="2025-10-02 10:06:07.677975886 +0000 UTC m=+1039.430617458" Oct 02 10:06:08 crc kubenswrapper[4934]: I1002 10:06:08.622482 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 02 10:06:08 crc kubenswrapper[4934]: I1002 10:06:08.649282 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d2b6e083-a523-48e2-8157-088036174e12","Type":"ContainerStarted","Data":"b2d1f1b04e17b195fbd97880fd7d7c42841c576a17071e6ee86a7e804f16c194"} Oct 02 10:06:08 crc kubenswrapper[4934]: I1002 10:06:08.649915 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:06:08 crc kubenswrapper[4934]: I1002 10:06:08.681125 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.481634217 podStartE2EDuration="52.681104748s" podCreationTimestamp="2025-10-02 10:05:16 +0000 UTC" firstStartedPulling="2025-10-02 10:05:18.26375361 +0000 UTC m=+990.016395132" lastFinishedPulling="2025-10-02 10:05:33.463224121 +0000 UTC m=+1005.215865663" observedRunningTime="2025-10-02 10:06:08.675163319 +0000 UTC m=+1040.427804861" watchObservedRunningTime="2025-10-02 10:06:08.681104748 +0000 UTC m=+1040.433746280" Oct 02 10:06:08 crc kubenswrapper[4934]: I1002 10:06:08.868748 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" Oct 02 10:06:08 crc kubenswrapper[4934]: I1002 10:06:08.925671 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-769b765785-hsdtk"] Oct 02 10:06:08 crc kubenswrapper[4934]: I1002 10:06:08.926081 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-769b765785-hsdtk" podUID="0302150f-a71a-45f0-a6b4-024664b21e51" containerName="dnsmasq-dns" containerID="cri-o://c021b99af4ef55634dacbbd18d21bee250069171cab0f36d2e5639562b6b4fa5" gracePeriod=10 Oct 02 10:06:09 crc kubenswrapper[4934]: I1002 10:06:09.395184 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-769b765785-hsdtk" Oct 02 10:06:09 crc kubenswrapper[4934]: I1002 10:06:09.564912 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0302150f-a71a-45f0-a6b4-024664b21e51-config\") pod \"0302150f-a71a-45f0-a6b4-024664b21e51\" (UID: \"0302150f-a71a-45f0-a6b4-024664b21e51\") " Oct 02 10:06:09 crc kubenswrapper[4934]: I1002 10:06:09.565033 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0302150f-a71a-45f0-a6b4-024664b21e51-dns-svc\") pod \"0302150f-a71a-45f0-a6b4-024664b21e51\" (UID: \"0302150f-a71a-45f0-a6b4-024664b21e51\") " Oct 02 10:06:09 crc kubenswrapper[4934]: I1002 10:06:09.565117 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ls99k\" (UniqueName: \"kubernetes.io/projected/0302150f-a71a-45f0-a6b4-024664b21e51-kube-api-access-ls99k\") pod \"0302150f-a71a-45f0-a6b4-024664b21e51\" (UID: \"0302150f-a71a-45f0-a6b4-024664b21e51\") " Oct 02 10:06:09 crc kubenswrapper[4934]: I1002 10:06:09.565137 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0302150f-a71a-45f0-a6b4-024664b21e51-ovsdbserver-nb\") pod \"0302150f-a71a-45f0-a6b4-024664b21e51\" (UID: \"0302150f-a71a-45f0-a6b4-024664b21e51\") " Oct 02 10:06:09 crc kubenswrapper[4934]: I1002 10:06:09.572341 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0302150f-a71a-45f0-a6b4-024664b21e51-kube-api-access-ls99k" (OuterVolumeSpecName: "kube-api-access-ls99k") pod "0302150f-a71a-45f0-a6b4-024664b21e51" (UID: "0302150f-a71a-45f0-a6b4-024664b21e51"). InnerVolumeSpecName "kube-api-access-ls99k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:09 crc kubenswrapper[4934]: I1002 10:06:09.603353 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0302150f-a71a-45f0-a6b4-024664b21e51-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0302150f-a71a-45f0-a6b4-024664b21e51" (UID: "0302150f-a71a-45f0-a6b4-024664b21e51"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:06:09 crc kubenswrapper[4934]: I1002 10:06:09.613079 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0302150f-a71a-45f0-a6b4-024664b21e51-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0302150f-a71a-45f0-a6b4-024664b21e51" (UID: "0302150f-a71a-45f0-a6b4-024664b21e51"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:06:09 crc kubenswrapper[4934]: I1002 10:06:09.616240 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0302150f-a71a-45f0-a6b4-024664b21e51-config" (OuterVolumeSpecName: "config") pod "0302150f-a71a-45f0-a6b4-024664b21e51" (UID: "0302150f-a71a-45f0-a6b4-024664b21e51"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:06:09 crc kubenswrapper[4934]: I1002 10:06:09.659938 4934 generic.go:334] "Generic (PLEG): container finished" podID="a1e933bf-e455-43c2-a996-b05a3bca3c96" containerID="7826f59cb7f15b3e25a5cbac82bed398ef57ad483fbbed25d37b37218b482cf8" exitCode=0 Oct 02 10:06:09 crc kubenswrapper[4934]: I1002 10:06:09.660850 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-n6t65" event={"ID":"a1e933bf-e455-43c2-a996-b05a3bca3c96","Type":"ContainerDied","Data":"7826f59cb7f15b3e25a5cbac82bed398ef57ad483fbbed25d37b37218b482cf8"} Oct 02 10:06:09 crc kubenswrapper[4934]: I1002 10:06:09.663886 4934 generic.go:334] "Generic (PLEG): container finished" podID="0302150f-a71a-45f0-a6b4-024664b21e51" containerID="c021b99af4ef55634dacbbd18d21bee250069171cab0f36d2e5639562b6b4fa5" exitCode=0 Oct 02 10:06:09 crc kubenswrapper[4934]: I1002 10:06:09.664834 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-769b765785-hsdtk" event={"ID":"0302150f-a71a-45f0-a6b4-024664b21e51","Type":"ContainerDied","Data":"c021b99af4ef55634dacbbd18d21bee250069171cab0f36d2e5639562b6b4fa5"} Oct 02 10:06:09 crc kubenswrapper[4934]: I1002 10:06:09.664883 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-769b765785-hsdtk" Oct 02 10:06:09 crc kubenswrapper[4934]: I1002 10:06:09.664918 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-769b765785-hsdtk" event={"ID":"0302150f-a71a-45f0-a6b4-024664b21e51","Type":"ContainerDied","Data":"705ad62eab9f24bd290443018ac7fc5b141723a5b8ef9cec64a7e76dd0402e31"} Oct 02 10:06:09 crc kubenswrapper[4934]: I1002 10:06:09.664947 4934 scope.go:117] "RemoveContainer" containerID="c021b99af4ef55634dacbbd18d21bee250069171cab0f36d2e5639562b6b4fa5" Oct 02 10:06:09 crc kubenswrapper[4934]: I1002 10:06:09.666724 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0302150f-a71a-45f0-a6b4-024664b21e51-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:09 crc kubenswrapper[4934]: I1002 10:06:09.666801 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ls99k\" (UniqueName: \"kubernetes.io/projected/0302150f-a71a-45f0-a6b4-024664b21e51-kube-api-access-ls99k\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:09 crc kubenswrapper[4934]: I1002 10:06:09.666873 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0302150f-a71a-45f0-a6b4-024664b21e51-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:09 crc kubenswrapper[4934]: I1002 10:06:09.667121 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0302150f-a71a-45f0-a6b4-024664b21e51-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:09 crc kubenswrapper[4934]: I1002 10:06:09.693375 4934 scope.go:117] "RemoveContainer" containerID="d5bf4f453290c95a9ed1c0b73753ce7ef8c66e50a522b798669bfe817dfa15a3" Oct 02 10:06:09 crc kubenswrapper[4934]: I1002 10:06:09.707250 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-769b765785-hsdtk"] Oct 02 10:06:09 crc kubenswrapper[4934]: I1002 10:06:09.711924 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-769b765785-hsdtk"] Oct 02 10:06:09 crc kubenswrapper[4934]: I1002 10:06:09.727588 4934 scope.go:117] "RemoveContainer" containerID="c021b99af4ef55634dacbbd18d21bee250069171cab0f36d2e5639562b6b4fa5" Oct 02 10:06:09 crc kubenswrapper[4934]: E1002 10:06:09.728134 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c021b99af4ef55634dacbbd18d21bee250069171cab0f36d2e5639562b6b4fa5\": container with ID starting with c021b99af4ef55634dacbbd18d21bee250069171cab0f36d2e5639562b6b4fa5 not found: ID does not exist" containerID="c021b99af4ef55634dacbbd18d21bee250069171cab0f36d2e5639562b6b4fa5" Oct 02 10:06:09 crc kubenswrapper[4934]: I1002 10:06:09.728216 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c021b99af4ef55634dacbbd18d21bee250069171cab0f36d2e5639562b6b4fa5"} err="failed to get container status \"c021b99af4ef55634dacbbd18d21bee250069171cab0f36d2e5639562b6b4fa5\": rpc error: code = NotFound desc = could not find container \"c021b99af4ef55634dacbbd18d21bee250069171cab0f36d2e5639562b6b4fa5\": container with ID starting with c021b99af4ef55634dacbbd18d21bee250069171cab0f36d2e5639562b6b4fa5 not found: ID does not exist" Oct 02 10:06:09 crc kubenswrapper[4934]: I1002 10:06:09.728289 4934 scope.go:117] "RemoveContainer" containerID="d5bf4f453290c95a9ed1c0b73753ce7ef8c66e50a522b798669bfe817dfa15a3" Oct 02 10:06:09 crc kubenswrapper[4934]: E1002 10:06:09.728728 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5bf4f453290c95a9ed1c0b73753ce7ef8c66e50a522b798669bfe817dfa15a3\": container with ID starting with d5bf4f453290c95a9ed1c0b73753ce7ef8c66e50a522b798669bfe817dfa15a3 not found: ID does not exist" containerID="d5bf4f453290c95a9ed1c0b73753ce7ef8c66e50a522b798669bfe817dfa15a3" Oct 02 10:06:09 crc kubenswrapper[4934]: I1002 10:06:09.728775 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5bf4f453290c95a9ed1c0b73753ce7ef8c66e50a522b798669bfe817dfa15a3"} err="failed to get container status \"d5bf4f453290c95a9ed1c0b73753ce7ef8c66e50a522b798669bfe817dfa15a3\": rpc error: code = NotFound desc = could not find container \"d5bf4f453290c95a9ed1c0b73753ce7ef8c66e50a522b798669bfe817dfa15a3\": container with ID starting with d5bf4f453290c95a9ed1c0b73753ce7ef8c66e50a522b798669bfe817dfa15a3 not found: ID does not exist" Oct 02 10:06:10 crc kubenswrapper[4934]: I1002 10:06:10.581745 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8cebd152-5aaf-4543-be11-7ba327cfdf0e-etc-swift\") pod \"swift-storage-0\" (UID: \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\") " pod="openstack/swift-storage-0" Oct 02 10:06:10 crc kubenswrapper[4934]: I1002 10:06:10.587196 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8cebd152-5aaf-4543-be11-7ba327cfdf0e-etc-swift\") pod \"swift-storage-0\" (UID: \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\") " pod="openstack/swift-storage-0" Oct 02 10:06:10 crc kubenswrapper[4934]: I1002 10:06:10.827683 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 02 10:06:10 crc kubenswrapper[4934]: I1002 10:06:10.931865 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0302150f-a71a-45f0-a6b4-024664b21e51" path="/var/lib/kubelet/pods/0302150f-a71a-45f0-a6b4-024664b21e51/volumes" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.007979 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-n6t65" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.090258 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a1e933bf-e455-43c2-a996-b05a3bca3c96-etc-swift\") pod \"a1e933bf-e455-43c2-a996-b05a3bca3c96\" (UID: \"a1e933bf-e455-43c2-a996-b05a3bca3c96\") " Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.090325 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1e933bf-e455-43c2-a996-b05a3bca3c96-scripts\") pod \"a1e933bf-e455-43c2-a996-b05a3bca3c96\" (UID: \"a1e933bf-e455-43c2-a996-b05a3bca3c96\") " Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.090427 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kchgx\" (UniqueName: \"kubernetes.io/projected/a1e933bf-e455-43c2-a996-b05a3bca3c96-kube-api-access-kchgx\") pod \"a1e933bf-e455-43c2-a996-b05a3bca3c96\" (UID: \"a1e933bf-e455-43c2-a996-b05a3bca3c96\") " Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.090482 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a1e933bf-e455-43c2-a996-b05a3bca3c96-ring-data-devices\") pod \"a1e933bf-e455-43c2-a996-b05a3bca3c96\" (UID: \"a1e933bf-e455-43c2-a996-b05a3bca3c96\") " Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.090520 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a1e933bf-e455-43c2-a996-b05a3bca3c96-dispersionconf\") pod \"a1e933bf-e455-43c2-a996-b05a3bca3c96\" (UID: \"a1e933bf-e455-43c2-a996-b05a3bca3c96\") " Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.090556 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1e933bf-e455-43c2-a996-b05a3bca3c96-combined-ca-bundle\") pod \"a1e933bf-e455-43c2-a996-b05a3bca3c96\" (UID: \"a1e933bf-e455-43c2-a996-b05a3bca3c96\") " Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.090677 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a1e933bf-e455-43c2-a996-b05a3bca3c96-swiftconf\") pod \"a1e933bf-e455-43c2-a996-b05a3bca3c96\" (UID: \"a1e933bf-e455-43c2-a996-b05a3bca3c96\") " Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.092145 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1e933bf-e455-43c2-a996-b05a3bca3c96-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "a1e933bf-e455-43c2-a996-b05a3bca3c96" (UID: "a1e933bf-e455-43c2-a996-b05a3bca3c96"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.092332 4934 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a1e933bf-e455-43c2-a996-b05a3bca3c96-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.093227 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1e933bf-e455-43c2-a996-b05a3bca3c96-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "a1e933bf-e455-43c2-a996-b05a3bca3c96" (UID: "a1e933bf-e455-43c2-a996-b05a3bca3c96"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.096658 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1e933bf-e455-43c2-a996-b05a3bca3c96-kube-api-access-kchgx" (OuterVolumeSpecName: "kube-api-access-kchgx") pod "a1e933bf-e455-43c2-a996-b05a3bca3c96" (UID: "a1e933bf-e455-43c2-a996-b05a3bca3c96"). InnerVolumeSpecName "kube-api-access-kchgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.105922 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1e933bf-e455-43c2-a996-b05a3bca3c96-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "a1e933bf-e455-43c2-a996-b05a3bca3c96" (UID: "a1e933bf-e455-43c2-a996-b05a3bca3c96"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.113855 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1e933bf-e455-43c2-a996-b05a3bca3c96-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "a1e933bf-e455-43c2-a996-b05a3bca3c96" (UID: "a1e933bf-e455-43c2-a996-b05a3bca3c96"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.126776 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1e933bf-e455-43c2-a996-b05a3bca3c96-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a1e933bf-e455-43c2-a996-b05a3bca3c96" (UID: "a1e933bf-e455-43c2-a996-b05a3bca3c96"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.131070 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1e933bf-e455-43c2-a996-b05a3bca3c96-scripts" (OuterVolumeSpecName: "scripts") pod "a1e933bf-e455-43c2-a996-b05a3bca3c96" (UID: "a1e933bf-e455-43c2-a996-b05a3bca3c96"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.184829 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-b73e-account-create-mnzcm"] Oct 02 10:06:11 crc kubenswrapper[4934]: E1002 10:06:11.185218 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1e933bf-e455-43c2-a996-b05a3bca3c96" containerName="swift-ring-rebalance" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.185241 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1e933bf-e455-43c2-a996-b05a3bca3c96" containerName="swift-ring-rebalance" Oct 02 10:06:11 crc kubenswrapper[4934]: E1002 10:06:11.185262 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0302150f-a71a-45f0-a6b4-024664b21e51" containerName="init" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.185271 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="0302150f-a71a-45f0-a6b4-024664b21e51" containerName="init" Oct 02 10:06:11 crc kubenswrapper[4934]: E1002 10:06:11.185288 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe286dcd-593e-48ee-8140-aca585425610" containerName="mariadb-database-create" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.185298 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe286dcd-593e-48ee-8140-aca585425610" containerName="mariadb-database-create" Oct 02 10:06:11 crc kubenswrapper[4934]: E1002 10:06:11.185309 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46721f65-4fa6-4059-aeb2-c9e62ed22cb1" containerName="mariadb-database-create" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.185319 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="46721f65-4fa6-4059-aeb2-c9e62ed22cb1" containerName="mariadb-database-create" Oct 02 10:06:11 crc kubenswrapper[4934]: E1002 10:06:11.185330 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0302150f-a71a-45f0-a6b4-024664b21e51" containerName="dnsmasq-dns" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.185337 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="0302150f-a71a-45f0-a6b4-024664b21e51" containerName="dnsmasq-dns" Oct 02 10:06:11 crc kubenswrapper[4934]: E1002 10:06:11.185359 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c705cde-0601-4f0e-97a3-590adbf0da4c" containerName="mariadb-database-create" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.185367 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c705cde-0601-4f0e-97a3-590adbf0da4c" containerName="mariadb-database-create" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.185543 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c705cde-0601-4f0e-97a3-590adbf0da4c" containerName="mariadb-database-create" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.185566 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="46721f65-4fa6-4059-aeb2-c9e62ed22cb1" containerName="mariadb-database-create" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.185593 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe286dcd-593e-48ee-8140-aca585425610" containerName="mariadb-database-create" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.185616 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1e933bf-e455-43c2-a996-b05a3bca3c96" containerName="swift-ring-rebalance" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.185625 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="0302150f-a71a-45f0-a6b4-024664b21e51" containerName="dnsmasq-dns" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.186270 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b73e-account-create-mnzcm" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.192828 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.194497 4934 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a1e933bf-e455-43c2-a996-b05a3bca3c96-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.194522 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1e933bf-e455-43c2-a996-b05a3bca3c96-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.194534 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kchgx\" (UniqueName: \"kubernetes.io/projected/a1e933bf-e455-43c2-a996-b05a3bca3c96-kube-api-access-kchgx\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.194546 4934 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a1e933bf-e455-43c2-a996-b05a3bca3c96-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.194556 4934 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a1e933bf-e455-43c2-a996-b05a3bca3c96-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.194566 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1e933bf-e455-43c2-a996-b05a3bca3c96-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.196836 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-b73e-account-create-mnzcm"] Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.296465 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnw5l\" (UniqueName: \"kubernetes.io/projected/f7fcfea2-ee1e-4ef2-ac7b-2948b0f289ac-kube-api-access-xnw5l\") pod \"keystone-b73e-account-create-mnzcm\" (UID: \"f7fcfea2-ee1e-4ef2-ac7b-2948b0f289ac\") " pod="openstack/keystone-b73e-account-create-mnzcm" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.398267 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnw5l\" (UniqueName: \"kubernetes.io/projected/f7fcfea2-ee1e-4ef2-ac7b-2948b0f289ac-kube-api-access-xnw5l\") pod \"keystone-b73e-account-create-mnzcm\" (UID: \"f7fcfea2-ee1e-4ef2-ac7b-2948b0f289ac\") " pod="openstack/keystone-b73e-account-create-mnzcm" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.415264 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnw5l\" (UniqueName: \"kubernetes.io/projected/f7fcfea2-ee1e-4ef2-ac7b-2948b0f289ac-kube-api-access-xnw5l\") pod \"keystone-b73e-account-create-mnzcm\" (UID: \"f7fcfea2-ee1e-4ef2-ac7b-2948b0f289ac\") " pod="openstack/keystone-b73e-account-create-mnzcm" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.447463 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.504963 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b73e-account-create-mnzcm" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.507167 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5c1f-account-create-z7zpn"] Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.508458 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5c1f-account-create-z7zpn" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.511550 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.518403 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5c1f-account-create-z7zpn"] Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.534987 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-gwpkm" podUID="7214291c-3c8e-4173-a6fd-ef16170258e2" containerName="ovn-controller" probeResult="failure" output=< Oct 02 10:06:11 crc kubenswrapper[4934]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 02 10:06:11 crc kubenswrapper[4934]: > Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.548783 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-65wtc" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.548855 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-65wtc" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.602759 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsbjx\" (UniqueName: \"kubernetes.io/projected/8841e9fd-caad-4ebd-8a89-659e10a0505b-kube-api-access-rsbjx\") pod \"placement-5c1f-account-create-z7zpn\" (UID: \"8841e9fd-caad-4ebd-8a89-659e10a0505b\") " pod="openstack/placement-5c1f-account-create-z7zpn" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.684934 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerStarted","Data":"44025b72bcf24f1079aba739bc296bb9f110537bfa5d3c6ec0f9d48d0c6b4a07"} Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.687744 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-n6t65" event={"ID":"a1e933bf-e455-43c2-a996-b05a3bca3c96","Type":"ContainerDied","Data":"618e539486eeeb7909a3130d7aef5c1e7f43783aabc80d44b99d43f4e73b6cf2"} Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.687784 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="618e539486eeeb7909a3130d7aef5c1e7f43783aabc80d44b99d43f4e73b6cf2" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.687807 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-n6t65" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.704129 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsbjx\" (UniqueName: \"kubernetes.io/projected/8841e9fd-caad-4ebd-8a89-659e10a0505b-kube-api-access-rsbjx\") pod \"placement-5c1f-account-create-z7zpn\" (UID: \"8841e9fd-caad-4ebd-8a89-659e10a0505b\") " pod="openstack/placement-5c1f-account-create-z7zpn" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.726483 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsbjx\" (UniqueName: \"kubernetes.io/projected/8841e9fd-caad-4ebd-8a89-659e10a0505b-kube-api-access-rsbjx\") pod \"placement-5c1f-account-create-z7zpn\" (UID: \"8841e9fd-caad-4ebd-8a89-659e10a0505b\") " pod="openstack/placement-5c1f-account-create-z7zpn" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.762731 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-gwpkm-config-dh8zg"] Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.763725 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gwpkm-config-dh8zg" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.768822 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.780305 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gwpkm-config-dh8zg"] Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.830250 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-b73e-account-create-mnzcm"] Oct 02 10:06:11 crc kubenswrapper[4934]: W1002 10:06:11.832260 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7fcfea2_ee1e_4ef2_ac7b_2948b0f289ac.slice/crio-e3a70f2559a7afbe556fcc89712a42248d39a0490a745962a5191d9afea5b737 WatchSource:0}: Error finding container e3a70f2559a7afbe556fcc89712a42248d39a0490a745962a5191d9afea5b737: Status 404 returned error can't find the container with id e3a70f2559a7afbe556fcc89712a42248d39a0490a745962a5191d9afea5b737 Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.841221 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5c1f-account-create-z7zpn" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.908020 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4079e146-8e07-4aa6-ba01-178d2d184c19-var-run-ovn\") pod \"ovn-controller-gwpkm-config-dh8zg\" (UID: \"4079e146-8e07-4aa6-ba01-178d2d184c19\") " pod="openstack/ovn-controller-gwpkm-config-dh8zg" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.908104 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4079e146-8e07-4aa6-ba01-178d2d184c19-additional-scripts\") pod \"ovn-controller-gwpkm-config-dh8zg\" (UID: \"4079e146-8e07-4aa6-ba01-178d2d184c19\") " pod="openstack/ovn-controller-gwpkm-config-dh8zg" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.908160 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7lnf\" (UniqueName: \"kubernetes.io/projected/4079e146-8e07-4aa6-ba01-178d2d184c19-kube-api-access-v7lnf\") pod \"ovn-controller-gwpkm-config-dh8zg\" (UID: \"4079e146-8e07-4aa6-ba01-178d2d184c19\") " pod="openstack/ovn-controller-gwpkm-config-dh8zg" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.908194 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4079e146-8e07-4aa6-ba01-178d2d184c19-var-run\") pod \"ovn-controller-gwpkm-config-dh8zg\" (UID: \"4079e146-8e07-4aa6-ba01-178d2d184c19\") " pod="openstack/ovn-controller-gwpkm-config-dh8zg" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.908312 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4079e146-8e07-4aa6-ba01-178d2d184c19-scripts\") pod \"ovn-controller-gwpkm-config-dh8zg\" (UID: \"4079e146-8e07-4aa6-ba01-178d2d184c19\") " pod="openstack/ovn-controller-gwpkm-config-dh8zg" Oct 02 10:06:11 crc kubenswrapper[4934]: I1002 10:06:11.908393 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4079e146-8e07-4aa6-ba01-178d2d184c19-var-log-ovn\") pod \"ovn-controller-gwpkm-config-dh8zg\" (UID: \"4079e146-8e07-4aa6-ba01-178d2d184c19\") " pod="openstack/ovn-controller-gwpkm-config-dh8zg" Oct 02 10:06:12 crc kubenswrapper[4934]: I1002 10:06:12.010268 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4079e146-8e07-4aa6-ba01-178d2d184c19-var-run-ovn\") pod \"ovn-controller-gwpkm-config-dh8zg\" (UID: \"4079e146-8e07-4aa6-ba01-178d2d184c19\") " pod="openstack/ovn-controller-gwpkm-config-dh8zg" Oct 02 10:06:12 crc kubenswrapper[4934]: I1002 10:06:12.010341 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4079e146-8e07-4aa6-ba01-178d2d184c19-additional-scripts\") pod \"ovn-controller-gwpkm-config-dh8zg\" (UID: \"4079e146-8e07-4aa6-ba01-178d2d184c19\") " pod="openstack/ovn-controller-gwpkm-config-dh8zg" Oct 02 10:06:12 crc kubenswrapper[4934]: I1002 10:06:12.010400 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7lnf\" (UniqueName: \"kubernetes.io/projected/4079e146-8e07-4aa6-ba01-178d2d184c19-kube-api-access-v7lnf\") pod \"ovn-controller-gwpkm-config-dh8zg\" (UID: \"4079e146-8e07-4aa6-ba01-178d2d184c19\") " pod="openstack/ovn-controller-gwpkm-config-dh8zg" Oct 02 10:06:12 crc kubenswrapper[4934]: I1002 10:06:12.010424 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4079e146-8e07-4aa6-ba01-178d2d184c19-var-run\") pod \"ovn-controller-gwpkm-config-dh8zg\" (UID: \"4079e146-8e07-4aa6-ba01-178d2d184c19\") " pod="openstack/ovn-controller-gwpkm-config-dh8zg" Oct 02 10:06:12 crc kubenswrapper[4934]: I1002 10:06:12.010442 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4079e146-8e07-4aa6-ba01-178d2d184c19-scripts\") pod \"ovn-controller-gwpkm-config-dh8zg\" (UID: \"4079e146-8e07-4aa6-ba01-178d2d184c19\") " pod="openstack/ovn-controller-gwpkm-config-dh8zg" Oct 02 10:06:12 crc kubenswrapper[4934]: I1002 10:06:12.010461 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4079e146-8e07-4aa6-ba01-178d2d184c19-var-log-ovn\") pod \"ovn-controller-gwpkm-config-dh8zg\" (UID: \"4079e146-8e07-4aa6-ba01-178d2d184c19\") " pod="openstack/ovn-controller-gwpkm-config-dh8zg" Oct 02 10:06:12 crc kubenswrapper[4934]: I1002 10:06:12.010590 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4079e146-8e07-4aa6-ba01-178d2d184c19-var-run-ovn\") pod \"ovn-controller-gwpkm-config-dh8zg\" (UID: \"4079e146-8e07-4aa6-ba01-178d2d184c19\") " pod="openstack/ovn-controller-gwpkm-config-dh8zg" Oct 02 10:06:12 crc kubenswrapper[4934]: I1002 10:06:12.010662 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4079e146-8e07-4aa6-ba01-178d2d184c19-var-run\") pod \"ovn-controller-gwpkm-config-dh8zg\" (UID: \"4079e146-8e07-4aa6-ba01-178d2d184c19\") " pod="openstack/ovn-controller-gwpkm-config-dh8zg" Oct 02 10:06:12 crc kubenswrapper[4934]: I1002 10:06:12.011568 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4079e146-8e07-4aa6-ba01-178d2d184c19-additional-scripts\") pod \"ovn-controller-gwpkm-config-dh8zg\" (UID: \"4079e146-8e07-4aa6-ba01-178d2d184c19\") " pod="openstack/ovn-controller-gwpkm-config-dh8zg" Oct 02 10:06:12 crc kubenswrapper[4934]: I1002 10:06:12.012287 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4079e146-8e07-4aa6-ba01-178d2d184c19-var-log-ovn\") pod \"ovn-controller-gwpkm-config-dh8zg\" (UID: \"4079e146-8e07-4aa6-ba01-178d2d184c19\") " pod="openstack/ovn-controller-gwpkm-config-dh8zg" Oct 02 10:06:12 crc kubenswrapper[4934]: I1002 10:06:12.013133 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4079e146-8e07-4aa6-ba01-178d2d184c19-scripts\") pod \"ovn-controller-gwpkm-config-dh8zg\" (UID: \"4079e146-8e07-4aa6-ba01-178d2d184c19\") " pod="openstack/ovn-controller-gwpkm-config-dh8zg" Oct 02 10:06:12 crc kubenswrapper[4934]: I1002 10:06:12.029857 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7lnf\" (UniqueName: \"kubernetes.io/projected/4079e146-8e07-4aa6-ba01-178d2d184c19-kube-api-access-v7lnf\") pod \"ovn-controller-gwpkm-config-dh8zg\" (UID: \"4079e146-8e07-4aa6-ba01-178d2d184c19\") " pod="openstack/ovn-controller-gwpkm-config-dh8zg" Oct 02 10:06:12 crc kubenswrapper[4934]: I1002 10:06:12.083936 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gwpkm-config-dh8zg" Oct 02 10:06:12 crc kubenswrapper[4934]: I1002 10:06:12.281241 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5c1f-account-create-z7zpn"] Oct 02 10:06:12 crc kubenswrapper[4934]: W1002 10:06:12.317120 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8841e9fd_caad_4ebd_8a89_659e10a0505b.slice/crio-663f2367e3e749a23f54231a8237255290b363fb9a437b2f474c9444faa613de WatchSource:0}: Error finding container 663f2367e3e749a23f54231a8237255290b363fb9a437b2f474c9444faa613de: Status 404 returned error can't find the container with id 663f2367e3e749a23f54231a8237255290b363fb9a437b2f474c9444faa613de Oct 02 10:06:12 crc kubenswrapper[4934]: I1002 10:06:12.389391 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gwpkm-config-dh8zg"] Oct 02 10:06:12 crc kubenswrapper[4934]: I1002 10:06:12.695962 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gwpkm-config-dh8zg" event={"ID":"4079e146-8e07-4aa6-ba01-178d2d184c19","Type":"ContainerStarted","Data":"3d9eacb74f023d7089e02655720751779d4405eaaa2eb536c60ce31055bcceba"} Oct 02 10:06:12 crc kubenswrapper[4934]: I1002 10:06:12.698722 4934 generic.go:334] "Generic (PLEG): container finished" podID="f7fcfea2-ee1e-4ef2-ac7b-2948b0f289ac" containerID="2ddbdd41bb6a64a519400ee4292c076412f2dc818cd3e1b0b3fc892184d900bb" exitCode=0 Oct 02 10:06:12 crc kubenswrapper[4934]: I1002 10:06:12.698784 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b73e-account-create-mnzcm" event={"ID":"f7fcfea2-ee1e-4ef2-ac7b-2948b0f289ac","Type":"ContainerDied","Data":"2ddbdd41bb6a64a519400ee4292c076412f2dc818cd3e1b0b3fc892184d900bb"} Oct 02 10:06:12 crc kubenswrapper[4934]: I1002 10:06:12.698811 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b73e-account-create-mnzcm" event={"ID":"f7fcfea2-ee1e-4ef2-ac7b-2948b0f289ac","Type":"ContainerStarted","Data":"e3a70f2559a7afbe556fcc89712a42248d39a0490a745962a5191d9afea5b737"} Oct 02 10:06:12 crc kubenswrapper[4934]: I1002 10:06:12.701023 4934 generic.go:334] "Generic (PLEG): container finished" podID="8841e9fd-caad-4ebd-8a89-659e10a0505b" containerID="0cda272c203bbc1f03d13a3985bd0215950af56f688bd7b5e079492809f581a7" exitCode=0 Oct 02 10:06:12 crc kubenswrapper[4934]: I1002 10:06:12.701047 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5c1f-account-create-z7zpn" event={"ID":"8841e9fd-caad-4ebd-8a89-659e10a0505b","Type":"ContainerDied","Data":"0cda272c203bbc1f03d13a3985bd0215950af56f688bd7b5e079492809f581a7"} Oct 02 10:06:12 crc kubenswrapper[4934]: I1002 10:06:12.701061 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5c1f-account-create-z7zpn" event={"ID":"8841e9fd-caad-4ebd-8a89-659e10a0505b","Type":"ContainerStarted","Data":"663f2367e3e749a23f54231a8237255290b363fb9a437b2f474c9444faa613de"} Oct 02 10:06:13 crc kubenswrapper[4934]: I1002 10:06:13.713156 4934 generic.go:334] "Generic (PLEG): container finished" podID="4079e146-8e07-4aa6-ba01-178d2d184c19" containerID="85cf8b87a7583c3d9bdd7099945fc472bfce2a71819309ed2cd711819382fe15" exitCode=0 Oct 02 10:06:13 crc kubenswrapper[4934]: I1002 10:06:13.713768 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gwpkm-config-dh8zg" event={"ID":"4079e146-8e07-4aa6-ba01-178d2d184c19","Type":"ContainerDied","Data":"85cf8b87a7583c3d9bdd7099945fc472bfce2a71819309ed2cd711819382fe15"} Oct 02 10:06:13 crc kubenswrapper[4934]: I1002 10:06:13.724541 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerStarted","Data":"65c0b15d452f5559c24817138f1d42f00b5d08d34eff5f6a036df1ded5c70ccb"} Oct 02 10:06:13 crc kubenswrapper[4934]: I1002 10:06:13.724833 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerStarted","Data":"5c785fb317467101f9ff407243fb85596511c5e525db3832b38df04ed00fec73"} Oct 02 10:06:13 crc kubenswrapper[4934]: I1002 10:06:13.725222 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerStarted","Data":"283ced1ad0dcd3181290d000eab3ac44a3033b8e32fa0ca0cd8ef03d57a7deee"} Oct 02 10:06:13 crc kubenswrapper[4934]: I1002 10:06:13.725440 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerStarted","Data":"478b442a459773165d46dac69a02032f32c9b37d6398eff28160ced5dd0ee4b4"} Oct 02 10:06:14 crc kubenswrapper[4934]: I1002 10:06:14.154907 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5c1f-account-create-z7zpn" Oct 02 10:06:14 crc kubenswrapper[4934]: I1002 10:06:14.160741 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b73e-account-create-mnzcm" Oct 02 10:06:14 crc kubenswrapper[4934]: I1002 10:06:14.260566 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rsbjx\" (UniqueName: \"kubernetes.io/projected/8841e9fd-caad-4ebd-8a89-659e10a0505b-kube-api-access-rsbjx\") pod \"8841e9fd-caad-4ebd-8a89-659e10a0505b\" (UID: \"8841e9fd-caad-4ebd-8a89-659e10a0505b\") " Oct 02 10:06:14 crc kubenswrapper[4934]: I1002 10:06:14.260764 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnw5l\" (UniqueName: \"kubernetes.io/projected/f7fcfea2-ee1e-4ef2-ac7b-2948b0f289ac-kube-api-access-xnw5l\") pod \"f7fcfea2-ee1e-4ef2-ac7b-2948b0f289ac\" (UID: \"f7fcfea2-ee1e-4ef2-ac7b-2948b0f289ac\") " Oct 02 10:06:14 crc kubenswrapper[4934]: I1002 10:06:14.265854 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7fcfea2-ee1e-4ef2-ac7b-2948b0f289ac-kube-api-access-xnw5l" (OuterVolumeSpecName: "kube-api-access-xnw5l") pod "f7fcfea2-ee1e-4ef2-ac7b-2948b0f289ac" (UID: "f7fcfea2-ee1e-4ef2-ac7b-2948b0f289ac"). InnerVolumeSpecName "kube-api-access-xnw5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:14 crc kubenswrapper[4934]: I1002 10:06:14.275732 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8841e9fd-caad-4ebd-8a89-659e10a0505b-kube-api-access-rsbjx" (OuterVolumeSpecName: "kube-api-access-rsbjx") pod "8841e9fd-caad-4ebd-8a89-659e10a0505b" (UID: "8841e9fd-caad-4ebd-8a89-659e10a0505b"). InnerVolumeSpecName "kube-api-access-rsbjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:14 crc kubenswrapper[4934]: I1002 10:06:14.362555 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnw5l\" (UniqueName: \"kubernetes.io/projected/f7fcfea2-ee1e-4ef2-ac7b-2948b0f289ac-kube-api-access-xnw5l\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:14 crc kubenswrapper[4934]: I1002 10:06:14.362598 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rsbjx\" (UniqueName: \"kubernetes.io/projected/8841e9fd-caad-4ebd-8a89-659e10a0505b-kube-api-access-rsbjx\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:14 crc kubenswrapper[4934]: I1002 10:06:14.736435 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b73e-account-create-mnzcm" event={"ID":"f7fcfea2-ee1e-4ef2-ac7b-2948b0f289ac","Type":"ContainerDied","Data":"e3a70f2559a7afbe556fcc89712a42248d39a0490a745962a5191d9afea5b737"} Oct 02 10:06:14 crc kubenswrapper[4934]: I1002 10:06:14.736482 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3a70f2559a7afbe556fcc89712a42248d39a0490a745962a5191d9afea5b737" Oct 02 10:06:14 crc kubenswrapper[4934]: I1002 10:06:14.736546 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b73e-account-create-mnzcm" Oct 02 10:06:14 crc kubenswrapper[4934]: I1002 10:06:14.743079 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5c1f-account-create-z7zpn" event={"ID":"8841e9fd-caad-4ebd-8a89-659e10a0505b","Type":"ContainerDied","Data":"663f2367e3e749a23f54231a8237255290b363fb9a437b2f474c9444faa613de"} Oct 02 10:06:14 crc kubenswrapper[4934]: I1002 10:06:14.743122 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="663f2367e3e749a23f54231a8237255290b363fb9a437b2f474c9444faa613de" Oct 02 10:06:14 crc kubenswrapper[4934]: I1002 10:06:14.743116 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5c1f-account-create-z7zpn" Oct 02 10:06:15 crc kubenswrapper[4934]: I1002 10:06:15.061706 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gwpkm-config-dh8zg" Oct 02 10:06:15 crc kubenswrapper[4934]: I1002 10:06:15.174558 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4079e146-8e07-4aa6-ba01-178d2d184c19-var-run\") pod \"4079e146-8e07-4aa6-ba01-178d2d184c19\" (UID: \"4079e146-8e07-4aa6-ba01-178d2d184c19\") " Oct 02 10:06:15 crc kubenswrapper[4934]: I1002 10:06:15.174643 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4079e146-8e07-4aa6-ba01-178d2d184c19-var-log-ovn\") pod \"4079e146-8e07-4aa6-ba01-178d2d184c19\" (UID: \"4079e146-8e07-4aa6-ba01-178d2d184c19\") " Oct 02 10:06:15 crc kubenswrapper[4934]: I1002 10:06:15.174721 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4079e146-8e07-4aa6-ba01-178d2d184c19-additional-scripts\") pod \"4079e146-8e07-4aa6-ba01-178d2d184c19\" (UID: \"4079e146-8e07-4aa6-ba01-178d2d184c19\") " Oct 02 10:06:15 crc kubenswrapper[4934]: I1002 10:06:15.174739 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4079e146-8e07-4aa6-ba01-178d2d184c19-var-run-ovn\") pod \"4079e146-8e07-4aa6-ba01-178d2d184c19\" (UID: \"4079e146-8e07-4aa6-ba01-178d2d184c19\") " Oct 02 10:06:15 crc kubenswrapper[4934]: I1002 10:06:15.174780 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7lnf\" (UniqueName: \"kubernetes.io/projected/4079e146-8e07-4aa6-ba01-178d2d184c19-kube-api-access-v7lnf\") pod \"4079e146-8e07-4aa6-ba01-178d2d184c19\" (UID: \"4079e146-8e07-4aa6-ba01-178d2d184c19\") " Oct 02 10:06:15 crc kubenswrapper[4934]: I1002 10:06:15.174815 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4079e146-8e07-4aa6-ba01-178d2d184c19-scripts\") pod \"4079e146-8e07-4aa6-ba01-178d2d184c19\" (UID: \"4079e146-8e07-4aa6-ba01-178d2d184c19\") " Oct 02 10:06:15 crc kubenswrapper[4934]: I1002 10:06:15.175487 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4079e146-8e07-4aa6-ba01-178d2d184c19-var-run" (OuterVolumeSpecName: "var-run") pod "4079e146-8e07-4aa6-ba01-178d2d184c19" (UID: "4079e146-8e07-4aa6-ba01-178d2d184c19"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:06:15 crc kubenswrapper[4934]: I1002 10:06:15.175520 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4079e146-8e07-4aa6-ba01-178d2d184c19-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "4079e146-8e07-4aa6-ba01-178d2d184c19" (UID: "4079e146-8e07-4aa6-ba01-178d2d184c19"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:06:15 crc kubenswrapper[4934]: I1002 10:06:15.176749 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4079e146-8e07-4aa6-ba01-178d2d184c19-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "4079e146-8e07-4aa6-ba01-178d2d184c19" (UID: "4079e146-8e07-4aa6-ba01-178d2d184c19"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:06:15 crc kubenswrapper[4934]: I1002 10:06:15.176890 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4079e146-8e07-4aa6-ba01-178d2d184c19-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "4079e146-8e07-4aa6-ba01-178d2d184c19" (UID: "4079e146-8e07-4aa6-ba01-178d2d184c19"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:06:15 crc kubenswrapper[4934]: I1002 10:06:15.177416 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4079e146-8e07-4aa6-ba01-178d2d184c19-scripts" (OuterVolumeSpecName: "scripts") pod "4079e146-8e07-4aa6-ba01-178d2d184c19" (UID: "4079e146-8e07-4aa6-ba01-178d2d184c19"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:06:15 crc kubenswrapper[4934]: I1002 10:06:15.199005 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4079e146-8e07-4aa6-ba01-178d2d184c19-kube-api-access-v7lnf" (OuterVolumeSpecName: "kube-api-access-v7lnf") pod "4079e146-8e07-4aa6-ba01-178d2d184c19" (UID: "4079e146-8e07-4aa6-ba01-178d2d184c19"). InnerVolumeSpecName "kube-api-access-v7lnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:15 crc kubenswrapper[4934]: I1002 10:06:15.276675 4934 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4079e146-8e07-4aa6-ba01-178d2d184c19-var-run\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:15 crc kubenswrapper[4934]: I1002 10:06:15.276699 4934 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4079e146-8e07-4aa6-ba01-178d2d184c19-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:15 crc kubenswrapper[4934]: I1002 10:06:15.276709 4934 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4079e146-8e07-4aa6-ba01-178d2d184c19-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:15 crc kubenswrapper[4934]: I1002 10:06:15.276718 4934 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4079e146-8e07-4aa6-ba01-178d2d184c19-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:15 crc kubenswrapper[4934]: I1002 10:06:15.276726 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7lnf\" (UniqueName: \"kubernetes.io/projected/4079e146-8e07-4aa6-ba01-178d2d184c19-kube-api-access-v7lnf\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:15 crc kubenswrapper[4934]: I1002 10:06:15.276734 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4079e146-8e07-4aa6-ba01-178d2d184c19-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:15 crc kubenswrapper[4934]: I1002 10:06:15.760021 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gwpkm-config-dh8zg" event={"ID":"4079e146-8e07-4aa6-ba01-178d2d184c19","Type":"ContainerDied","Data":"3d9eacb74f023d7089e02655720751779d4405eaaa2eb536c60ce31055bcceba"} Oct 02 10:06:15 crc kubenswrapper[4934]: I1002 10:06:15.760366 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d9eacb74f023d7089e02655720751779d4405eaaa2eb536c60ce31055bcceba" Oct 02 10:06:15 crc kubenswrapper[4934]: I1002 10:06:15.760210 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gwpkm-config-dh8zg" Oct 02 10:06:15 crc kubenswrapper[4934]: I1002 10:06:15.765743 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerStarted","Data":"883824276a85ba5f915880ccf065542294a5a3959ead2ee9e02e556f4e13a641"} Oct 02 10:06:15 crc kubenswrapper[4934]: I1002 10:06:15.765782 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerStarted","Data":"022023c76cc2ba5ec65353c6ec2387833901f691565a07dc55b9a6a09bf9871c"} Oct 02 10:06:15 crc kubenswrapper[4934]: I1002 10:06:15.765793 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerStarted","Data":"da203f48b1e4a0c44c371939bd0a1b6ee7159680a2a3fac865d9476c3b7b9c05"} Oct 02 10:06:15 crc kubenswrapper[4934]: I1002 10:06:15.765800 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerStarted","Data":"c0f5815d02eff4beefc969cb8468e9b85a60d871f99cc29f33aa27f2cd50039f"} Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.169663 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-gwpkm-config-dh8zg"] Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.176251 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-gwpkm-config-dh8zg"] Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.268784 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-gwpkm-config-fh72z"] Oct 02 10:06:16 crc kubenswrapper[4934]: E1002 10:06:16.269210 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4079e146-8e07-4aa6-ba01-178d2d184c19" containerName="ovn-config" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.269234 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="4079e146-8e07-4aa6-ba01-178d2d184c19" containerName="ovn-config" Oct 02 10:06:16 crc kubenswrapper[4934]: E1002 10:06:16.269252 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8841e9fd-caad-4ebd-8a89-659e10a0505b" containerName="mariadb-account-create" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.269260 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8841e9fd-caad-4ebd-8a89-659e10a0505b" containerName="mariadb-account-create" Oct 02 10:06:16 crc kubenswrapper[4934]: E1002 10:06:16.269286 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7fcfea2-ee1e-4ef2-ac7b-2948b0f289ac" containerName="mariadb-account-create" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.269293 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7fcfea2-ee1e-4ef2-ac7b-2948b0f289ac" containerName="mariadb-account-create" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.269500 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="4079e146-8e07-4aa6-ba01-178d2d184c19" containerName="ovn-config" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.269552 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="8841e9fd-caad-4ebd-8a89-659e10a0505b" containerName="mariadb-account-create" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.269575 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7fcfea2-ee1e-4ef2-ac7b-2948b0f289ac" containerName="mariadb-account-create" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.270137 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gwpkm-config-fh72z" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.272651 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.283610 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gwpkm-config-fh72z"] Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.393217 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0d8af8e2-931f-451a-91fe-9cb0751028b6-var-run\") pod \"ovn-controller-gwpkm-config-fh72z\" (UID: \"0d8af8e2-931f-451a-91fe-9cb0751028b6\") " pod="openstack/ovn-controller-gwpkm-config-fh72z" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.393269 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0d8af8e2-931f-451a-91fe-9cb0751028b6-additional-scripts\") pod \"ovn-controller-gwpkm-config-fh72z\" (UID: \"0d8af8e2-931f-451a-91fe-9cb0751028b6\") " pod="openstack/ovn-controller-gwpkm-config-fh72z" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.393334 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sks4c\" (UniqueName: \"kubernetes.io/projected/0d8af8e2-931f-451a-91fe-9cb0751028b6-kube-api-access-sks4c\") pod \"ovn-controller-gwpkm-config-fh72z\" (UID: \"0d8af8e2-931f-451a-91fe-9cb0751028b6\") " pod="openstack/ovn-controller-gwpkm-config-fh72z" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.393524 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0d8af8e2-931f-451a-91fe-9cb0751028b6-var-run-ovn\") pod \"ovn-controller-gwpkm-config-fh72z\" (UID: \"0d8af8e2-931f-451a-91fe-9cb0751028b6\") " pod="openstack/ovn-controller-gwpkm-config-fh72z" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.393702 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0d8af8e2-931f-451a-91fe-9cb0751028b6-var-log-ovn\") pod \"ovn-controller-gwpkm-config-fh72z\" (UID: \"0d8af8e2-931f-451a-91fe-9cb0751028b6\") " pod="openstack/ovn-controller-gwpkm-config-fh72z" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.393785 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0d8af8e2-931f-451a-91fe-9cb0751028b6-scripts\") pod \"ovn-controller-gwpkm-config-fh72z\" (UID: \"0d8af8e2-931f-451a-91fe-9cb0751028b6\") " pod="openstack/ovn-controller-gwpkm-config-fh72z" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.497055 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0d8af8e2-931f-451a-91fe-9cb0751028b6-var-run-ovn\") pod \"ovn-controller-gwpkm-config-fh72z\" (UID: \"0d8af8e2-931f-451a-91fe-9cb0751028b6\") " pod="openstack/ovn-controller-gwpkm-config-fh72z" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.497201 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0d8af8e2-931f-451a-91fe-9cb0751028b6-var-log-ovn\") pod \"ovn-controller-gwpkm-config-fh72z\" (UID: \"0d8af8e2-931f-451a-91fe-9cb0751028b6\") " pod="openstack/ovn-controller-gwpkm-config-fh72z" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.497350 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0d8af8e2-931f-451a-91fe-9cb0751028b6-var-run-ovn\") pod \"ovn-controller-gwpkm-config-fh72z\" (UID: \"0d8af8e2-931f-451a-91fe-9cb0751028b6\") " pod="openstack/ovn-controller-gwpkm-config-fh72z" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.497319 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0d8af8e2-931f-451a-91fe-9cb0751028b6-scripts\") pod \"ovn-controller-gwpkm-config-fh72z\" (UID: \"0d8af8e2-931f-451a-91fe-9cb0751028b6\") " pod="openstack/ovn-controller-gwpkm-config-fh72z" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.497451 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0d8af8e2-931f-451a-91fe-9cb0751028b6-var-log-ovn\") pod \"ovn-controller-gwpkm-config-fh72z\" (UID: \"0d8af8e2-931f-451a-91fe-9cb0751028b6\") " pod="openstack/ovn-controller-gwpkm-config-fh72z" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.497653 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0d8af8e2-931f-451a-91fe-9cb0751028b6-var-run\") pod \"ovn-controller-gwpkm-config-fh72z\" (UID: \"0d8af8e2-931f-451a-91fe-9cb0751028b6\") " pod="openstack/ovn-controller-gwpkm-config-fh72z" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.497720 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0d8af8e2-931f-451a-91fe-9cb0751028b6-additional-scripts\") pod \"ovn-controller-gwpkm-config-fh72z\" (UID: \"0d8af8e2-931f-451a-91fe-9cb0751028b6\") " pod="openstack/ovn-controller-gwpkm-config-fh72z" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.497831 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0d8af8e2-931f-451a-91fe-9cb0751028b6-var-run\") pod \"ovn-controller-gwpkm-config-fh72z\" (UID: \"0d8af8e2-931f-451a-91fe-9cb0751028b6\") " pod="openstack/ovn-controller-gwpkm-config-fh72z" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.497941 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sks4c\" (UniqueName: \"kubernetes.io/projected/0d8af8e2-931f-451a-91fe-9cb0751028b6-kube-api-access-sks4c\") pod \"ovn-controller-gwpkm-config-fh72z\" (UID: \"0d8af8e2-931f-451a-91fe-9cb0751028b6\") " pod="openstack/ovn-controller-gwpkm-config-fh72z" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.499082 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0d8af8e2-931f-451a-91fe-9cb0751028b6-additional-scripts\") pod \"ovn-controller-gwpkm-config-fh72z\" (UID: \"0d8af8e2-931f-451a-91fe-9cb0751028b6\") " pod="openstack/ovn-controller-gwpkm-config-fh72z" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.504175 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0d8af8e2-931f-451a-91fe-9cb0751028b6-scripts\") pod \"ovn-controller-gwpkm-config-fh72z\" (UID: \"0d8af8e2-931f-451a-91fe-9cb0751028b6\") " pod="openstack/ovn-controller-gwpkm-config-fh72z" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.534127 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sks4c\" (UniqueName: \"kubernetes.io/projected/0d8af8e2-931f-451a-91fe-9cb0751028b6-kube-api-access-sks4c\") pod \"ovn-controller-gwpkm-config-fh72z\" (UID: \"0d8af8e2-931f-451a-91fe-9cb0751028b6\") " pod="openstack/ovn-controller-gwpkm-config-fh72z" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.542520 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-gwpkm" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.594281 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gwpkm-config-fh72z" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.792185 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-82a2-account-create-nd4xr"] Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.793226 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-82a2-account-create-nd4xr" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.796048 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.809934 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-82a2-account-create-nd4xr"] Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.908830 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dg66b\" (UniqueName: \"kubernetes.io/projected/1c63f790-0724-4a40-8b22-9895eb3b7f8b-kube-api-access-dg66b\") pod \"glance-82a2-account-create-nd4xr\" (UID: \"1c63f790-0724-4a40-8b22-9895eb3b7f8b\") " pod="openstack/glance-82a2-account-create-nd4xr" Oct 02 10:06:16 crc kubenswrapper[4934]: I1002 10:06:16.929900 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4079e146-8e07-4aa6-ba01-178d2d184c19" path="/var/lib/kubelet/pods/4079e146-8e07-4aa6-ba01-178d2d184c19/volumes" Oct 02 10:06:17 crc kubenswrapper[4934]: I1002 10:06:17.010414 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dg66b\" (UniqueName: \"kubernetes.io/projected/1c63f790-0724-4a40-8b22-9895eb3b7f8b-kube-api-access-dg66b\") pod \"glance-82a2-account-create-nd4xr\" (UID: \"1c63f790-0724-4a40-8b22-9895eb3b7f8b\") " pod="openstack/glance-82a2-account-create-nd4xr" Oct 02 10:06:17 crc kubenswrapper[4934]: I1002 10:06:17.034558 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dg66b\" (UniqueName: \"kubernetes.io/projected/1c63f790-0724-4a40-8b22-9895eb3b7f8b-kube-api-access-dg66b\") pod \"glance-82a2-account-create-nd4xr\" (UID: \"1c63f790-0724-4a40-8b22-9895eb3b7f8b\") " pod="openstack/glance-82a2-account-create-nd4xr" Oct 02 10:06:17 crc kubenswrapper[4934]: I1002 10:06:17.047674 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gwpkm-config-fh72z"] Oct 02 10:06:17 crc kubenswrapper[4934]: I1002 10:06:17.116817 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-82a2-account-create-nd4xr" Oct 02 10:06:17 crc kubenswrapper[4934]: W1002 10:06:17.209737 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d8af8e2_931f_451a_91fe_9cb0751028b6.slice/crio-5d415c86aafd7d5a956da330f7424161bc255b413babc274e81f804c9848f048 WatchSource:0}: Error finding container 5d415c86aafd7d5a956da330f7424161bc255b413babc274e81f804c9848f048: Status 404 returned error can't find the container with id 5d415c86aafd7d5a956da330f7424161bc255b413babc274e81f804c9848f048 Oct 02 10:06:17 crc kubenswrapper[4934]: I1002 10:06:17.658885 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-82a2-account-create-nd4xr"] Oct 02 10:06:17 crc kubenswrapper[4934]: I1002 10:06:17.812639 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gwpkm-config-fh72z" event={"ID":"0d8af8e2-931f-451a-91fe-9cb0751028b6","Type":"ContainerStarted","Data":"5d6e15890d9c5182fa81723619291a5028257eb0f92591a5c0724ba0f008bbbd"} Oct 02 10:06:17 crc kubenswrapper[4934]: I1002 10:06:17.812676 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gwpkm-config-fh72z" event={"ID":"0d8af8e2-931f-451a-91fe-9cb0751028b6","Type":"ContainerStarted","Data":"5d415c86aafd7d5a956da330f7424161bc255b413babc274e81f804c9848f048"} Oct 02 10:06:17 crc kubenswrapper[4934]: I1002 10:06:17.817224 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-82a2-account-create-nd4xr" event={"ID":"1c63f790-0724-4a40-8b22-9895eb3b7f8b","Type":"ContainerStarted","Data":"44cbc58904357ccb74b7a52e8de377efe6ad49fc61491217ae01eae39dc23381"} Oct 02 10:06:17 crc kubenswrapper[4934]: I1002 10:06:17.822355 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerStarted","Data":"d6fc4c846f604ef5014de443285ee7dab569b41691babd5a6981585d8007df21"} Oct 02 10:06:17 crc kubenswrapper[4934]: I1002 10:06:17.822397 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerStarted","Data":"1e2427a94989f9a77ca59046a739fc7849881116c29e79c766ddc52debf75ec3"} Oct 02 10:06:17 crc kubenswrapper[4934]: I1002 10:06:17.830879 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-gwpkm-config-fh72z" podStartSLOduration=1.830854103 podStartE2EDuration="1.830854103s" podCreationTimestamp="2025-10-02 10:06:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:06:17.828529877 +0000 UTC m=+1049.581171389" watchObservedRunningTime="2025-10-02 10:06:17.830854103 +0000 UTC m=+1049.583495635" Oct 02 10:06:17 crc kubenswrapper[4934]: I1002 10:06:17.850955 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:06:18 crc kubenswrapper[4934]: I1002 10:06:18.626338 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 02 10:06:18 crc kubenswrapper[4934]: I1002 10:06:18.832498 4934 generic.go:334] "Generic (PLEG): container finished" podID="0d8af8e2-931f-451a-91fe-9cb0751028b6" containerID="5d6e15890d9c5182fa81723619291a5028257eb0f92591a5c0724ba0f008bbbd" exitCode=0 Oct 02 10:06:18 crc kubenswrapper[4934]: I1002 10:06:18.833158 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gwpkm-config-fh72z" event={"ID":"0d8af8e2-931f-451a-91fe-9cb0751028b6","Type":"ContainerDied","Data":"5d6e15890d9c5182fa81723619291a5028257eb0f92591a5c0724ba0f008bbbd"} Oct 02 10:06:18 crc kubenswrapper[4934]: I1002 10:06:18.834281 4934 generic.go:334] "Generic (PLEG): container finished" podID="1c63f790-0724-4a40-8b22-9895eb3b7f8b" containerID="d90d083e790725945807012f42289291311c29e6d6bd45cb9bdc6528aa2de2e4" exitCode=0 Oct 02 10:06:18 crc kubenswrapper[4934]: I1002 10:06:18.834337 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-82a2-account-create-nd4xr" event={"ID":"1c63f790-0724-4a40-8b22-9895eb3b7f8b","Type":"ContainerDied","Data":"d90d083e790725945807012f42289291311c29e6d6bd45cb9bdc6528aa2de2e4"} Oct 02 10:06:18 crc kubenswrapper[4934]: I1002 10:06:18.840514 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerStarted","Data":"9373e5fbc3b166030467efc3d48ed6860fef2654fa4d21f813440b32765b3063"} Oct 02 10:06:18 crc kubenswrapper[4934]: I1002 10:06:18.840550 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerStarted","Data":"2d1cebd6bce0741d006ac48e38db4c872771c02036472a9fa71980ddfb72cef5"} Oct 02 10:06:18 crc kubenswrapper[4934]: I1002 10:06:18.840559 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerStarted","Data":"6b7e46bf6361363ea3ce0effc8f79e7c186a4d64e12c8053c444d0f89313722e"} Oct 02 10:06:18 crc kubenswrapper[4934]: I1002 10:06:18.840568 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerStarted","Data":"169e61c78b12ac121ef1306849a0fe5e05cc7125ba6f6a238a652aac3445bad8"} Oct 02 10:06:19 crc kubenswrapper[4934]: I1002 10:06:19.783696 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-sj7h4"] Oct 02 10:06:19 crc kubenswrapper[4934]: I1002 10:06:19.785053 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-sj7h4" Oct 02 10:06:19 crc kubenswrapper[4934]: I1002 10:06:19.792660 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-sj7h4"] Oct 02 10:06:19 crc kubenswrapper[4934]: I1002 10:06:19.855414 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerStarted","Data":"1c65da253bd0c434afbafb0cd05ae9a507db27a1ff5ce1626446a181ca57c304"} Oct 02 10:06:19 crc kubenswrapper[4934]: I1002 10:06:19.861467 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqgc7\" (UniqueName: \"kubernetes.io/projected/52c52806-3d2a-4fac-938d-ba4a58ba77d5-kube-api-access-vqgc7\") pod \"cinder-db-create-sj7h4\" (UID: \"52c52806-3d2a-4fac-938d-ba4a58ba77d5\") " pod="openstack/cinder-db-create-sj7h4" Oct 02 10:06:19 crc kubenswrapper[4934]: I1002 10:06:19.895774 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-qm5fm"] Oct 02 10:06:19 crc kubenswrapper[4934]: I1002 10:06:19.897093 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-qm5fm" Oct 02 10:06:19 crc kubenswrapper[4934]: I1002 10:06:19.913878 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=21.106711664 podStartE2EDuration="26.913860055s" podCreationTimestamp="2025-10-02 10:05:53 +0000 UTC" firstStartedPulling="2025-10-02 10:06:11.467677604 +0000 UTC m=+1043.220319126" lastFinishedPulling="2025-10-02 10:06:17.274825995 +0000 UTC m=+1049.027467517" observedRunningTime="2025-10-02 10:06:19.905650531 +0000 UTC m=+1051.658292063" watchObservedRunningTime="2025-10-02 10:06:19.913860055 +0000 UTC m=+1051.666501577" Oct 02 10:06:19 crc kubenswrapper[4934]: I1002 10:06:19.946189 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-qm5fm"] Oct 02 10:06:19 crc kubenswrapper[4934]: I1002 10:06:19.962766 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqgc7\" (UniqueName: \"kubernetes.io/projected/52c52806-3d2a-4fac-938d-ba4a58ba77d5-kube-api-access-vqgc7\") pod \"cinder-db-create-sj7h4\" (UID: \"52c52806-3d2a-4fac-938d-ba4a58ba77d5\") " pod="openstack/cinder-db-create-sj7h4" Oct 02 10:06:19 crc kubenswrapper[4934]: I1002 10:06:19.962875 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwhg5\" (UniqueName: \"kubernetes.io/projected/ae0d619d-6fbb-4758-80bc-95bdfe6f2022-kube-api-access-cwhg5\") pod \"barbican-db-create-qm5fm\" (UID: \"ae0d619d-6fbb-4758-80bc-95bdfe6f2022\") " pod="openstack/barbican-db-create-qm5fm" Oct 02 10:06:19 crc kubenswrapper[4934]: I1002 10:06:19.993474 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqgc7\" (UniqueName: \"kubernetes.io/projected/52c52806-3d2a-4fac-938d-ba4a58ba77d5-kube-api-access-vqgc7\") pod \"cinder-db-create-sj7h4\" (UID: \"52c52806-3d2a-4fac-938d-ba4a58ba77d5\") " pod="openstack/cinder-db-create-sj7h4" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.064840 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwhg5\" (UniqueName: \"kubernetes.io/projected/ae0d619d-6fbb-4758-80bc-95bdfe6f2022-kube-api-access-cwhg5\") pod \"barbican-db-create-qm5fm\" (UID: \"ae0d619d-6fbb-4758-80bc-95bdfe6f2022\") " pod="openstack/barbican-db-create-qm5fm" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.089407 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-cw4qq"] Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.091367 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwhg5\" (UniqueName: \"kubernetes.io/projected/ae0d619d-6fbb-4758-80bc-95bdfe6f2022-kube-api-access-cwhg5\") pod \"barbican-db-create-qm5fm\" (UID: \"ae0d619d-6fbb-4758-80bc-95bdfe6f2022\") " pod="openstack/barbican-db-create-qm5fm" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.097759 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-cw4qq" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.104025 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-sj7h4" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.120155 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-cw4qq"] Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.166737 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hmkl\" (UniqueName: \"kubernetes.io/projected/30903f5b-46ed-4b59-ad1e-073466cac566-kube-api-access-8hmkl\") pod \"neutron-db-create-cw4qq\" (UID: \"30903f5b-46ed-4b59-ad1e-073466cac566\") " pod="openstack/neutron-db-create-cw4qq" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.221516 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-qm5fm" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.236616 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-c25xv"] Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.238720 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-c25xv" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.263649 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-bkm4l" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.264223 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.271848 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f4da5aa-1b82-4286-8a3b-68ca8f486c78-config-data\") pod \"keystone-db-sync-c25xv\" (UID: \"6f4da5aa-1b82-4286-8a3b-68ca8f486c78\") " pod="openstack/keystone-db-sync-c25xv" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.271995 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f4da5aa-1b82-4286-8a3b-68ca8f486c78-combined-ca-bundle\") pod \"keystone-db-sync-c25xv\" (UID: \"6f4da5aa-1b82-4286-8a3b-68ca8f486c78\") " pod="openstack/keystone-db-sync-c25xv" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.272060 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hmkl\" (UniqueName: \"kubernetes.io/projected/30903f5b-46ed-4b59-ad1e-073466cac566-kube-api-access-8hmkl\") pod \"neutron-db-create-cw4qq\" (UID: \"30903f5b-46ed-4b59-ad1e-073466cac566\") " pod="openstack/neutron-db-create-cw4qq" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.272090 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66gz8\" (UniqueName: \"kubernetes.io/projected/6f4da5aa-1b82-4286-8a3b-68ca8f486c78-kube-api-access-66gz8\") pod \"keystone-db-sync-c25xv\" (UID: \"6f4da5aa-1b82-4286-8a3b-68ca8f486c78\") " pod="openstack/keystone-db-sync-c25xv" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.282192 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.282364 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.282445 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-c25xv"] Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.309218 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hmkl\" (UniqueName: \"kubernetes.io/projected/30903f5b-46ed-4b59-ad1e-073466cac566-kube-api-access-8hmkl\") pod \"neutron-db-create-cw4qq\" (UID: \"30903f5b-46ed-4b59-ad1e-073466cac566\") " pod="openstack/neutron-db-create-cw4qq" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.317228 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-68d47f7449-5tf7r"] Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.319568 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.323900 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.335361 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68d47f7449-5tf7r"] Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.373489 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-ovsdbserver-sb\") pod \"dnsmasq-dns-68d47f7449-5tf7r\" (UID: \"95d46109-8486-474d-8b03-78088500da28\") " pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.373530 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-ovsdbserver-nb\") pod \"dnsmasq-dns-68d47f7449-5tf7r\" (UID: \"95d46109-8486-474d-8b03-78088500da28\") " pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.373571 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f4da5aa-1b82-4286-8a3b-68ca8f486c78-combined-ca-bundle\") pod \"keystone-db-sync-c25xv\" (UID: \"6f4da5aa-1b82-4286-8a3b-68ca8f486c78\") " pod="openstack/keystone-db-sync-c25xv" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.373636 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-82a2-account-create-nd4xr" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.373641 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66gz8\" (UniqueName: \"kubernetes.io/projected/6f4da5aa-1b82-4286-8a3b-68ca8f486c78-kube-api-access-66gz8\") pod \"keystone-db-sync-c25xv\" (UID: \"6f4da5aa-1b82-4286-8a3b-68ca8f486c78\") " pod="openstack/keystone-db-sync-c25xv" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.373927 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-config\") pod \"dnsmasq-dns-68d47f7449-5tf7r\" (UID: \"95d46109-8486-474d-8b03-78088500da28\") " pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.374078 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trnj8\" (UniqueName: \"kubernetes.io/projected/95d46109-8486-474d-8b03-78088500da28-kube-api-access-trnj8\") pod \"dnsmasq-dns-68d47f7449-5tf7r\" (UID: \"95d46109-8486-474d-8b03-78088500da28\") " pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.374107 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f4da5aa-1b82-4286-8a3b-68ca8f486c78-config-data\") pod \"keystone-db-sync-c25xv\" (UID: \"6f4da5aa-1b82-4286-8a3b-68ca8f486c78\") " pod="openstack/keystone-db-sync-c25xv" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.374137 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-dns-swift-storage-0\") pod \"dnsmasq-dns-68d47f7449-5tf7r\" (UID: \"95d46109-8486-474d-8b03-78088500da28\") " pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.374177 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-dns-svc\") pod \"dnsmasq-dns-68d47f7449-5tf7r\" (UID: \"95d46109-8486-474d-8b03-78088500da28\") " pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.383742 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f4da5aa-1b82-4286-8a3b-68ca8f486c78-combined-ca-bundle\") pod \"keystone-db-sync-c25xv\" (UID: \"6f4da5aa-1b82-4286-8a3b-68ca8f486c78\") " pod="openstack/keystone-db-sync-c25xv" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.397031 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gwpkm-config-fh72z" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.400601 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66gz8\" (UniqueName: \"kubernetes.io/projected/6f4da5aa-1b82-4286-8a3b-68ca8f486c78-kube-api-access-66gz8\") pod \"keystone-db-sync-c25xv\" (UID: \"6f4da5aa-1b82-4286-8a3b-68ca8f486c78\") " pod="openstack/keystone-db-sync-c25xv" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.410526 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f4da5aa-1b82-4286-8a3b-68ca8f486c78-config-data\") pod \"keystone-db-sync-c25xv\" (UID: \"6f4da5aa-1b82-4286-8a3b-68ca8f486c78\") " pod="openstack/keystone-db-sync-c25xv" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.475769 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0d8af8e2-931f-451a-91fe-9cb0751028b6-scripts\") pod \"0d8af8e2-931f-451a-91fe-9cb0751028b6\" (UID: \"0d8af8e2-931f-451a-91fe-9cb0751028b6\") " Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.475843 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0d8af8e2-931f-451a-91fe-9cb0751028b6-var-log-ovn\") pod \"0d8af8e2-931f-451a-91fe-9cb0751028b6\" (UID: \"0d8af8e2-931f-451a-91fe-9cb0751028b6\") " Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.475918 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0d8af8e2-931f-451a-91fe-9cb0751028b6-var-run\") pod \"0d8af8e2-931f-451a-91fe-9cb0751028b6\" (UID: \"0d8af8e2-931f-451a-91fe-9cb0751028b6\") " Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.475936 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0d8af8e2-931f-451a-91fe-9cb0751028b6-var-run-ovn\") pod \"0d8af8e2-931f-451a-91fe-9cb0751028b6\" (UID: \"0d8af8e2-931f-451a-91fe-9cb0751028b6\") " Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.476032 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dg66b\" (UniqueName: \"kubernetes.io/projected/1c63f790-0724-4a40-8b22-9895eb3b7f8b-kube-api-access-dg66b\") pod \"1c63f790-0724-4a40-8b22-9895eb3b7f8b\" (UID: \"1c63f790-0724-4a40-8b22-9895eb3b7f8b\") " Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.476098 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sks4c\" (UniqueName: \"kubernetes.io/projected/0d8af8e2-931f-451a-91fe-9cb0751028b6-kube-api-access-sks4c\") pod \"0d8af8e2-931f-451a-91fe-9cb0751028b6\" (UID: \"0d8af8e2-931f-451a-91fe-9cb0751028b6\") " Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.476120 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0d8af8e2-931f-451a-91fe-9cb0751028b6-additional-scripts\") pod \"0d8af8e2-931f-451a-91fe-9cb0751028b6\" (UID: \"0d8af8e2-931f-451a-91fe-9cb0751028b6\") " Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.476333 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-config\") pod \"dnsmasq-dns-68d47f7449-5tf7r\" (UID: \"95d46109-8486-474d-8b03-78088500da28\") " pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.476382 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trnj8\" (UniqueName: \"kubernetes.io/projected/95d46109-8486-474d-8b03-78088500da28-kube-api-access-trnj8\") pod \"dnsmasq-dns-68d47f7449-5tf7r\" (UID: \"95d46109-8486-474d-8b03-78088500da28\") " pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.476428 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-dns-swift-storage-0\") pod \"dnsmasq-dns-68d47f7449-5tf7r\" (UID: \"95d46109-8486-474d-8b03-78088500da28\") " pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.476483 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-dns-svc\") pod \"dnsmasq-dns-68d47f7449-5tf7r\" (UID: \"95d46109-8486-474d-8b03-78088500da28\") " pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.476517 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-ovsdbserver-sb\") pod \"dnsmasq-dns-68d47f7449-5tf7r\" (UID: \"95d46109-8486-474d-8b03-78088500da28\") " pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.476542 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-ovsdbserver-nb\") pod \"dnsmasq-dns-68d47f7449-5tf7r\" (UID: \"95d46109-8486-474d-8b03-78088500da28\") " pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.477258 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0d8af8e2-931f-451a-91fe-9cb0751028b6-var-run" (OuterVolumeSpecName: "var-run") pod "0d8af8e2-931f-451a-91fe-9cb0751028b6" (UID: "0d8af8e2-931f-451a-91fe-9cb0751028b6"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.478122 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d8af8e2-931f-451a-91fe-9cb0751028b6-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "0d8af8e2-931f-451a-91fe-9cb0751028b6" (UID: "0d8af8e2-931f-451a-91fe-9cb0751028b6"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.478357 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d8af8e2-931f-451a-91fe-9cb0751028b6-scripts" (OuterVolumeSpecName: "scripts") pod "0d8af8e2-931f-451a-91fe-9cb0751028b6" (UID: "0d8af8e2-931f-451a-91fe-9cb0751028b6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.478404 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0d8af8e2-931f-451a-91fe-9cb0751028b6-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "0d8af8e2-931f-451a-91fe-9cb0751028b6" (UID: "0d8af8e2-931f-451a-91fe-9cb0751028b6"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.478420 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-ovsdbserver-nb\") pod \"dnsmasq-dns-68d47f7449-5tf7r\" (UID: \"95d46109-8486-474d-8b03-78088500da28\") " pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.478471 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0d8af8e2-931f-451a-91fe-9cb0751028b6-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "0d8af8e2-931f-451a-91fe-9cb0751028b6" (UID: "0d8af8e2-931f-451a-91fe-9cb0751028b6"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.479545 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-config\") pod \"dnsmasq-dns-68d47f7449-5tf7r\" (UID: \"95d46109-8486-474d-8b03-78088500da28\") " pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.480105 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-dns-svc\") pod \"dnsmasq-dns-68d47f7449-5tf7r\" (UID: \"95d46109-8486-474d-8b03-78088500da28\") " pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.480277 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d8af8e2-931f-451a-91fe-9cb0751028b6-kube-api-access-sks4c" (OuterVolumeSpecName: "kube-api-access-sks4c") pod "0d8af8e2-931f-451a-91fe-9cb0751028b6" (UID: "0d8af8e2-931f-451a-91fe-9cb0751028b6"). InnerVolumeSpecName "kube-api-access-sks4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.480837 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-ovsdbserver-sb\") pod \"dnsmasq-dns-68d47f7449-5tf7r\" (UID: \"95d46109-8486-474d-8b03-78088500da28\") " pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.480975 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-dns-swift-storage-0\") pod \"dnsmasq-dns-68d47f7449-5tf7r\" (UID: \"95d46109-8486-474d-8b03-78088500da28\") " pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.482702 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c63f790-0724-4a40-8b22-9895eb3b7f8b-kube-api-access-dg66b" (OuterVolumeSpecName: "kube-api-access-dg66b") pod "1c63f790-0724-4a40-8b22-9895eb3b7f8b" (UID: "1c63f790-0724-4a40-8b22-9895eb3b7f8b"). InnerVolumeSpecName "kube-api-access-dg66b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.498712 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trnj8\" (UniqueName: \"kubernetes.io/projected/95d46109-8486-474d-8b03-78088500da28-kube-api-access-trnj8\") pod \"dnsmasq-dns-68d47f7449-5tf7r\" (UID: \"95d46109-8486-474d-8b03-78088500da28\") " pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.538803 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-cw4qq" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.578531 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0d8af8e2-931f-451a-91fe-9cb0751028b6-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.578567 4934 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0d8af8e2-931f-451a-91fe-9cb0751028b6-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.578593 4934 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0d8af8e2-931f-451a-91fe-9cb0751028b6-var-run\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.578602 4934 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0d8af8e2-931f-451a-91fe-9cb0751028b6-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.578611 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dg66b\" (UniqueName: \"kubernetes.io/projected/1c63f790-0724-4a40-8b22-9895eb3b7f8b-kube-api-access-dg66b\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.578620 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sks4c\" (UniqueName: \"kubernetes.io/projected/0d8af8e2-931f-451a-91fe-9cb0751028b6-kube-api-access-sks4c\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.578628 4934 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0d8af8e2-931f-451a-91fe-9cb0751028b6-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.614299 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-c25xv" Oct 02 10:06:20 crc kubenswrapper[4934]: I1002 10:06:20.658529 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:20.744161 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-sj7h4"] Oct 02 10:06:21 crc kubenswrapper[4934]: W1002 10:06:20.755003 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52c52806_3d2a_4fac_938d_ba4a58ba77d5.slice/crio-75185be0b2cd3acb1c3bd5a3050e36dd06c4ac40560e287764cd80ec9e1a4e5a WatchSource:0}: Error finding container 75185be0b2cd3acb1c3bd5a3050e36dd06c4ac40560e287764cd80ec9e1a4e5a: Status 404 returned error can't find the container with id 75185be0b2cd3acb1c3bd5a3050e36dd06c4ac40560e287764cd80ec9e1a4e5a Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:20.786888 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-qm5fm"] Oct 02 10:06:21 crc kubenswrapper[4934]: W1002 10:06:20.804031 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae0d619d_6fbb_4758_80bc_95bdfe6f2022.slice/crio-d2c33af521b91006806501de550f1e5bd2f4d34a127de4d053fa5946ad778633 WatchSource:0}: Error finding container d2c33af521b91006806501de550f1e5bd2f4d34a127de4d053fa5946ad778633: Status 404 returned error can't find the container with id d2c33af521b91006806501de550f1e5bd2f4d34a127de4d053fa5946ad778633 Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:20.871375 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gwpkm-config-fh72z" event={"ID":"0d8af8e2-931f-451a-91fe-9cb0751028b6","Type":"ContainerDied","Data":"5d415c86aafd7d5a956da330f7424161bc255b413babc274e81f804c9848f048"} Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:20.871692 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d415c86aafd7d5a956da330f7424161bc255b413babc274e81f804c9848f048" Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:20.871438 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gwpkm-config-fh72z" Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:20.872432 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-qm5fm" event={"ID":"ae0d619d-6fbb-4758-80bc-95bdfe6f2022","Type":"ContainerStarted","Data":"d2c33af521b91006806501de550f1e5bd2f4d34a127de4d053fa5946ad778633"} Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:20.894786 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-82a2-account-create-nd4xr" event={"ID":"1c63f790-0724-4a40-8b22-9895eb3b7f8b","Type":"ContainerDied","Data":"44cbc58904357ccb74b7a52e8de377efe6ad49fc61491217ae01eae39dc23381"} Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:20.894820 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44cbc58904357ccb74b7a52e8de377efe6ad49fc61491217ae01eae39dc23381" Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:20.894876 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-82a2-account-create-nd4xr" Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:20.902075 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-sj7h4" event={"ID":"52c52806-3d2a-4fac-938d-ba4a58ba77d5","Type":"ContainerStarted","Data":"75185be0b2cd3acb1c3bd5a3050e36dd06c4ac40560e287764cd80ec9e1a4e5a"} Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:20.928397 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-gwpkm-config-fh72z"] Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:20.932126 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-gwpkm-config-fh72z"] Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:21.869351 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-jwpkw"] Oct 02 10:06:21 crc kubenswrapper[4934]: E1002 10:06:21.870324 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c63f790-0724-4a40-8b22-9895eb3b7f8b" containerName="mariadb-account-create" Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:21.870342 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c63f790-0724-4a40-8b22-9895eb3b7f8b" containerName="mariadb-account-create" Oct 02 10:06:21 crc kubenswrapper[4934]: E1002 10:06:21.870366 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d8af8e2-931f-451a-91fe-9cb0751028b6" containerName="ovn-config" Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:21.870374 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d8af8e2-931f-451a-91fe-9cb0751028b6" containerName="ovn-config" Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:21.870571 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d8af8e2-931f-451a-91fe-9cb0751028b6" containerName="ovn-config" Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:21.870612 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c63f790-0724-4a40-8b22-9895eb3b7f8b" containerName="mariadb-account-create" Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:21.871286 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jwpkw" Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:21.874382 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:21.874617 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-6f9q2" Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:21.879209 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-jwpkw"] Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:21.897675 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6427286d-a1ca-43ca-8ecb-bbdda45e360e-config-data\") pod \"glance-db-sync-jwpkw\" (UID: \"6427286d-a1ca-43ca-8ecb-bbdda45e360e\") " pod="openstack/glance-db-sync-jwpkw" Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:21.897719 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6427286d-a1ca-43ca-8ecb-bbdda45e360e-combined-ca-bundle\") pod \"glance-db-sync-jwpkw\" (UID: \"6427286d-a1ca-43ca-8ecb-bbdda45e360e\") " pod="openstack/glance-db-sync-jwpkw" Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:21.897742 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6427286d-a1ca-43ca-8ecb-bbdda45e360e-db-sync-config-data\") pod \"glance-db-sync-jwpkw\" (UID: \"6427286d-a1ca-43ca-8ecb-bbdda45e360e\") " pod="openstack/glance-db-sync-jwpkw" Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:21.897781 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqq9t\" (UniqueName: \"kubernetes.io/projected/6427286d-a1ca-43ca-8ecb-bbdda45e360e-kube-api-access-zqq9t\") pod \"glance-db-sync-jwpkw\" (UID: \"6427286d-a1ca-43ca-8ecb-bbdda45e360e\") " pod="openstack/glance-db-sync-jwpkw" Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:21.911636 4934 generic.go:334] "Generic (PLEG): container finished" podID="52c52806-3d2a-4fac-938d-ba4a58ba77d5" containerID="17fa067adca99d11768b4048a418152c7da822434414f68329addaa440085afb" exitCode=0 Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:21.911718 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-sj7h4" event={"ID":"52c52806-3d2a-4fac-938d-ba4a58ba77d5","Type":"ContainerDied","Data":"17fa067adca99d11768b4048a418152c7da822434414f68329addaa440085afb"} Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:21.913312 4934 generic.go:334] "Generic (PLEG): container finished" podID="ae0d619d-6fbb-4758-80bc-95bdfe6f2022" containerID="742b3375416bacd7728d2de60efffabe7d330310983b7154b9fa6d01154d25d2" exitCode=0 Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:21.913342 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-qm5fm" event={"ID":"ae0d619d-6fbb-4758-80bc-95bdfe6f2022","Type":"ContainerDied","Data":"742b3375416bacd7728d2de60efffabe7d330310983b7154b9fa6d01154d25d2"} Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:21.968613 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-cw4qq"] Oct 02 10:06:21 crc kubenswrapper[4934]: I1002 10:06:21.999229 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6427286d-a1ca-43ca-8ecb-bbdda45e360e-config-data\") pod \"glance-db-sync-jwpkw\" (UID: \"6427286d-a1ca-43ca-8ecb-bbdda45e360e\") " pod="openstack/glance-db-sync-jwpkw" Oct 02 10:06:22 crc kubenswrapper[4934]: I1002 10:06:22.005057 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6427286d-a1ca-43ca-8ecb-bbdda45e360e-combined-ca-bundle\") pod \"glance-db-sync-jwpkw\" (UID: \"6427286d-a1ca-43ca-8ecb-bbdda45e360e\") " pod="openstack/glance-db-sync-jwpkw" Oct 02 10:06:22 crc kubenswrapper[4934]: I1002 10:06:22.004655 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6427286d-a1ca-43ca-8ecb-bbdda45e360e-config-data\") pod \"glance-db-sync-jwpkw\" (UID: \"6427286d-a1ca-43ca-8ecb-bbdda45e360e\") " pod="openstack/glance-db-sync-jwpkw" Oct 02 10:06:22 crc kubenswrapper[4934]: I1002 10:06:22.005129 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6427286d-a1ca-43ca-8ecb-bbdda45e360e-db-sync-config-data\") pod \"glance-db-sync-jwpkw\" (UID: \"6427286d-a1ca-43ca-8ecb-bbdda45e360e\") " pod="openstack/glance-db-sync-jwpkw" Oct 02 10:06:22 crc kubenswrapper[4934]: I1002 10:06:22.005210 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqq9t\" (UniqueName: \"kubernetes.io/projected/6427286d-a1ca-43ca-8ecb-bbdda45e360e-kube-api-access-zqq9t\") pod \"glance-db-sync-jwpkw\" (UID: \"6427286d-a1ca-43ca-8ecb-bbdda45e360e\") " pod="openstack/glance-db-sync-jwpkw" Oct 02 10:06:22 crc kubenswrapper[4934]: I1002 10:06:22.018433 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6427286d-a1ca-43ca-8ecb-bbdda45e360e-combined-ca-bundle\") pod \"glance-db-sync-jwpkw\" (UID: \"6427286d-a1ca-43ca-8ecb-bbdda45e360e\") " pod="openstack/glance-db-sync-jwpkw" Oct 02 10:06:22 crc kubenswrapper[4934]: I1002 10:06:22.018710 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6427286d-a1ca-43ca-8ecb-bbdda45e360e-db-sync-config-data\") pod \"glance-db-sync-jwpkw\" (UID: \"6427286d-a1ca-43ca-8ecb-bbdda45e360e\") " pod="openstack/glance-db-sync-jwpkw" Oct 02 10:06:22 crc kubenswrapper[4934]: I1002 10:06:22.024894 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqq9t\" (UniqueName: \"kubernetes.io/projected/6427286d-a1ca-43ca-8ecb-bbdda45e360e-kube-api-access-zqq9t\") pod \"glance-db-sync-jwpkw\" (UID: \"6427286d-a1ca-43ca-8ecb-bbdda45e360e\") " pod="openstack/glance-db-sync-jwpkw" Oct 02 10:06:22 crc kubenswrapper[4934]: I1002 10:06:22.050343 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-c25xv"] Oct 02 10:06:22 crc kubenswrapper[4934]: W1002 10:06:22.056252 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f4da5aa_1b82_4286_8a3b_68ca8f486c78.slice/crio-8f3db57be78c368ad20e5871f54d698c79fe3ee4af0f516f759204c87c225a64 WatchSource:0}: Error finding container 8f3db57be78c368ad20e5871f54d698c79fe3ee4af0f516f759204c87c225a64: Status 404 returned error can't find the container with id 8f3db57be78c368ad20e5871f54d698c79fe3ee4af0f516f759204c87c225a64 Oct 02 10:06:22 crc kubenswrapper[4934]: I1002 10:06:22.058926 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68d47f7449-5tf7r"] Oct 02 10:06:22 crc kubenswrapper[4934]: I1002 10:06:22.188943 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jwpkw" Oct 02 10:06:22 crc kubenswrapper[4934]: I1002 10:06:22.718522 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-jwpkw"] Oct 02 10:06:22 crc kubenswrapper[4934]: I1002 10:06:22.922909 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d8af8e2-931f-451a-91fe-9cb0751028b6" path="/var/lib/kubelet/pods/0d8af8e2-931f-451a-91fe-9cb0751028b6/volumes" Oct 02 10:06:22 crc kubenswrapper[4934]: I1002 10:06:22.923702 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-c25xv" event={"ID":"6f4da5aa-1b82-4286-8a3b-68ca8f486c78","Type":"ContainerStarted","Data":"8f3db57be78c368ad20e5871f54d698c79fe3ee4af0f516f759204c87c225a64"} Oct 02 10:06:22 crc kubenswrapper[4934]: I1002 10:06:22.925429 4934 generic.go:334] "Generic (PLEG): container finished" podID="95d46109-8486-474d-8b03-78088500da28" containerID="97c0006b77d839328731430ad86d47bb8ca5d18e833d3d3a511d11adab1fd1df" exitCode=0 Oct 02 10:06:22 crc kubenswrapper[4934]: I1002 10:06:22.925481 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" event={"ID":"95d46109-8486-474d-8b03-78088500da28","Type":"ContainerDied","Data":"97c0006b77d839328731430ad86d47bb8ca5d18e833d3d3a511d11adab1fd1df"} Oct 02 10:06:22 crc kubenswrapper[4934]: I1002 10:06:22.925531 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" event={"ID":"95d46109-8486-474d-8b03-78088500da28","Type":"ContainerStarted","Data":"5ba0bc49fe537e10fe256fd5be21a999522b603a1bb47fbfe679c6f97ee5fafa"} Oct 02 10:06:22 crc kubenswrapper[4934]: I1002 10:06:22.928833 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jwpkw" event={"ID":"6427286d-a1ca-43ca-8ecb-bbdda45e360e","Type":"ContainerStarted","Data":"35c7791c313ed28fdfce6e3b778e7be233cc1a891db6dfa945b93e0eb921a6a6"} Oct 02 10:06:22 crc kubenswrapper[4934]: I1002 10:06:22.933815 4934 generic.go:334] "Generic (PLEG): container finished" podID="30903f5b-46ed-4b59-ad1e-073466cac566" containerID="15c244cad2ff64e103eac7d91d2b751cb8d73809c1159f6386ea5b15b4dae2c9" exitCode=0 Oct 02 10:06:22 crc kubenswrapper[4934]: I1002 10:06:22.933877 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-cw4qq" event={"ID":"30903f5b-46ed-4b59-ad1e-073466cac566","Type":"ContainerDied","Data":"15c244cad2ff64e103eac7d91d2b751cb8d73809c1159f6386ea5b15b4dae2c9"} Oct 02 10:06:22 crc kubenswrapper[4934]: I1002 10:06:22.933943 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-cw4qq" event={"ID":"30903f5b-46ed-4b59-ad1e-073466cac566","Type":"ContainerStarted","Data":"fd9e18c6825562a35db84c99ccd5c6821dc7359a5f4aea963612adbf97b37c93"} Oct 02 10:06:23 crc kubenswrapper[4934]: I1002 10:06:23.335334 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-qm5fm" Oct 02 10:06:23 crc kubenswrapper[4934]: I1002 10:06:23.343063 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-sj7h4" Oct 02 10:06:23 crc kubenswrapper[4934]: I1002 10:06:23.428639 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqgc7\" (UniqueName: \"kubernetes.io/projected/52c52806-3d2a-4fac-938d-ba4a58ba77d5-kube-api-access-vqgc7\") pod \"52c52806-3d2a-4fac-938d-ba4a58ba77d5\" (UID: \"52c52806-3d2a-4fac-938d-ba4a58ba77d5\") " Oct 02 10:06:23 crc kubenswrapper[4934]: I1002 10:06:23.428939 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwhg5\" (UniqueName: \"kubernetes.io/projected/ae0d619d-6fbb-4758-80bc-95bdfe6f2022-kube-api-access-cwhg5\") pod \"ae0d619d-6fbb-4758-80bc-95bdfe6f2022\" (UID: \"ae0d619d-6fbb-4758-80bc-95bdfe6f2022\") " Oct 02 10:06:23 crc kubenswrapper[4934]: I1002 10:06:23.435744 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52c52806-3d2a-4fac-938d-ba4a58ba77d5-kube-api-access-vqgc7" (OuterVolumeSpecName: "kube-api-access-vqgc7") pod "52c52806-3d2a-4fac-938d-ba4a58ba77d5" (UID: "52c52806-3d2a-4fac-938d-ba4a58ba77d5"). InnerVolumeSpecName "kube-api-access-vqgc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:23 crc kubenswrapper[4934]: I1002 10:06:23.448172 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae0d619d-6fbb-4758-80bc-95bdfe6f2022-kube-api-access-cwhg5" (OuterVolumeSpecName: "kube-api-access-cwhg5") pod "ae0d619d-6fbb-4758-80bc-95bdfe6f2022" (UID: "ae0d619d-6fbb-4758-80bc-95bdfe6f2022"). InnerVolumeSpecName "kube-api-access-cwhg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:23 crc kubenswrapper[4934]: I1002 10:06:23.531665 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwhg5\" (UniqueName: \"kubernetes.io/projected/ae0d619d-6fbb-4758-80bc-95bdfe6f2022-kube-api-access-cwhg5\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:23 crc kubenswrapper[4934]: I1002 10:06:23.531710 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqgc7\" (UniqueName: \"kubernetes.io/projected/52c52806-3d2a-4fac-938d-ba4a58ba77d5-kube-api-access-vqgc7\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:23 crc kubenswrapper[4934]: I1002 10:06:23.948531 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-sj7h4" event={"ID":"52c52806-3d2a-4fac-938d-ba4a58ba77d5","Type":"ContainerDied","Data":"75185be0b2cd3acb1c3bd5a3050e36dd06c4ac40560e287764cd80ec9e1a4e5a"} Oct 02 10:06:23 crc kubenswrapper[4934]: I1002 10:06:23.948958 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75185be0b2cd3acb1c3bd5a3050e36dd06c4ac40560e287764cd80ec9e1a4e5a" Oct 02 10:06:23 crc kubenswrapper[4934]: I1002 10:06:23.948624 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-sj7h4" Oct 02 10:06:23 crc kubenswrapper[4934]: I1002 10:06:23.952251 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" event={"ID":"95d46109-8486-474d-8b03-78088500da28","Type":"ContainerStarted","Data":"e20b9a73c9c60067b166a4179b8e24057a49922f4c4019066a179535b25fa21c"} Oct 02 10:06:23 crc kubenswrapper[4934]: I1002 10:06:23.952907 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" Oct 02 10:06:23 crc kubenswrapper[4934]: I1002 10:06:23.957208 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-qm5fm" event={"ID":"ae0d619d-6fbb-4758-80bc-95bdfe6f2022","Type":"ContainerDied","Data":"d2c33af521b91006806501de550f1e5bd2f4d34a127de4d053fa5946ad778633"} Oct 02 10:06:23 crc kubenswrapper[4934]: I1002 10:06:23.957235 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-qm5fm" Oct 02 10:06:23 crc kubenswrapper[4934]: I1002 10:06:23.957244 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2c33af521b91006806501de550f1e5bd2f4d34a127de4d053fa5946ad778633" Oct 02 10:06:23 crc kubenswrapper[4934]: I1002 10:06:23.973443 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" podStartSLOduration=3.973427485 podStartE2EDuration="3.973427485s" podCreationTimestamp="2025-10-02 10:06:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:06:23.97009181 +0000 UTC m=+1055.722733352" watchObservedRunningTime="2025-10-02 10:06:23.973427485 +0000 UTC m=+1055.726069007" Oct 02 10:06:24 crc kubenswrapper[4934]: E1002 10:06:24.061494 4934 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52c52806_3d2a_4fac_938d_ba4a58ba77d5.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae0d619d_6fbb_4758_80bc_95bdfe6f2022.slice\": RecentStats: unable to find data in memory cache]" Oct 02 10:06:26 crc kubenswrapper[4934]: I1002 10:06:26.804510 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-cw4qq" Oct 02 10:06:26 crc kubenswrapper[4934]: I1002 10:06:26.888319 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hmkl\" (UniqueName: \"kubernetes.io/projected/30903f5b-46ed-4b59-ad1e-073466cac566-kube-api-access-8hmkl\") pod \"30903f5b-46ed-4b59-ad1e-073466cac566\" (UID: \"30903f5b-46ed-4b59-ad1e-073466cac566\") " Oct 02 10:06:26 crc kubenswrapper[4934]: I1002 10:06:26.893296 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30903f5b-46ed-4b59-ad1e-073466cac566-kube-api-access-8hmkl" (OuterVolumeSpecName: "kube-api-access-8hmkl") pod "30903f5b-46ed-4b59-ad1e-073466cac566" (UID: "30903f5b-46ed-4b59-ad1e-073466cac566"). InnerVolumeSpecName "kube-api-access-8hmkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:26 crc kubenswrapper[4934]: I1002 10:06:26.985012 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-cw4qq" event={"ID":"30903f5b-46ed-4b59-ad1e-073466cac566","Type":"ContainerDied","Data":"fd9e18c6825562a35db84c99ccd5c6821dc7359a5f4aea963612adbf97b37c93"} Oct 02 10:06:26 crc kubenswrapper[4934]: I1002 10:06:26.985055 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd9e18c6825562a35db84c99ccd5c6821dc7359a5f4aea963612adbf97b37c93" Oct 02 10:06:26 crc kubenswrapper[4934]: I1002 10:06:26.985073 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-cw4qq" Oct 02 10:06:26 crc kubenswrapper[4934]: I1002 10:06:26.991893 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hmkl\" (UniqueName: \"kubernetes.io/projected/30903f5b-46ed-4b59-ad1e-073466cac566-kube-api-access-8hmkl\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:27 crc kubenswrapper[4934]: I1002 10:06:27.996538 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-c25xv" event={"ID":"6f4da5aa-1b82-4286-8a3b-68ca8f486c78","Type":"ContainerStarted","Data":"8075fe4eb85efceb29e57e7b750c07c1739933c65104eec3beb30a995332f635"} Oct 02 10:06:28 crc kubenswrapper[4934]: I1002 10:06:28.016915 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-c25xv" podStartSLOduration=3.25821136 podStartE2EDuration="8.016896766s" podCreationTimestamp="2025-10-02 10:06:20 +0000 UTC" firstStartedPulling="2025-10-02 10:06:22.058800183 +0000 UTC m=+1053.811441705" lastFinishedPulling="2025-10-02 10:06:26.817485589 +0000 UTC m=+1058.570127111" observedRunningTime="2025-10-02 10:06:28.013926351 +0000 UTC m=+1059.766567873" watchObservedRunningTime="2025-10-02 10:06:28.016896766 +0000 UTC m=+1059.769538288" Oct 02 10:06:29 crc kubenswrapper[4934]: I1002 10:06:29.915110 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-36fe-account-create-8rqjp"] Oct 02 10:06:29 crc kubenswrapper[4934]: E1002 10:06:29.915475 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52c52806-3d2a-4fac-938d-ba4a58ba77d5" containerName="mariadb-database-create" Oct 02 10:06:29 crc kubenswrapper[4934]: I1002 10:06:29.915491 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="52c52806-3d2a-4fac-938d-ba4a58ba77d5" containerName="mariadb-database-create" Oct 02 10:06:29 crc kubenswrapper[4934]: E1002 10:06:29.915513 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae0d619d-6fbb-4758-80bc-95bdfe6f2022" containerName="mariadb-database-create" Oct 02 10:06:29 crc kubenswrapper[4934]: I1002 10:06:29.915521 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae0d619d-6fbb-4758-80bc-95bdfe6f2022" containerName="mariadb-database-create" Oct 02 10:06:29 crc kubenswrapper[4934]: E1002 10:06:29.915552 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30903f5b-46ed-4b59-ad1e-073466cac566" containerName="mariadb-database-create" Oct 02 10:06:29 crc kubenswrapper[4934]: I1002 10:06:29.915561 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="30903f5b-46ed-4b59-ad1e-073466cac566" containerName="mariadb-database-create" Oct 02 10:06:29 crc kubenswrapper[4934]: I1002 10:06:29.915781 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae0d619d-6fbb-4758-80bc-95bdfe6f2022" containerName="mariadb-database-create" Oct 02 10:06:29 crc kubenswrapper[4934]: I1002 10:06:29.916041 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="30903f5b-46ed-4b59-ad1e-073466cac566" containerName="mariadb-database-create" Oct 02 10:06:29 crc kubenswrapper[4934]: I1002 10:06:29.916052 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="52c52806-3d2a-4fac-938d-ba4a58ba77d5" containerName="mariadb-database-create" Oct 02 10:06:29 crc kubenswrapper[4934]: I1002 10:06:29.916792 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-36fe-account-create-8rqjp" Oct 02 10:06:29 crc kubenswrapper[4934]: I1002 10:06:29.920202 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 02 10:06:29 crc kubenswrapper[4934]: I1002 10:06:29.924052 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-36fe-account-create-8rqjp"] Oct 02 10:06:29 crc kubenswrapper[4934]: I1002 10:06:29.948683 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kdxj\" (UniqueName: \"kubernetes.io/projected/be42c16a-682c-44b2-a53a-d33085d3a681-kube-api-access-2kdxj\") pod \"cinder-36fe-account-create-8rqjp\" (UID: \"be42c16a-682c-44b2-a53a-d33085d3a681\") " pod="openstack/cinder-36fe-account-create-8rqjp" Oct 02 10:06:30 crc kubenswrapper[4934]: I1002 10:06:30.050851 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kdxj\" (UniqueName: \"kubernetes.io/projected/be42c16a-682c-44b2-a53a-d33085d3a681-kube-api-access-2kdxj\") pod \"cinder-36fe-account-create-8rqjp\" (UID: \"be42c16a-682c-44b2-a53a-d33085d3a681\") " pod="openstack/cinder-36fe-account-create-8rqjp" Oct 02 10:06:30 crc kubenswrapper[4934]: I1002 10:06:30.077354 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kdxj\" (UniqueName: \"kubernetes.io/projected/be42c16a-682c-44b2-a53a-d33085d3a681-kube-api-access-2kdxj\") pod \"cinder-36fe-account-create-8rqjp\" (UID: \"be42c16a-682c-44b2-a53a-d33085d3a681\") " pod="openstack/cinder-36fe-account-create-8rqjp" Oct 02 10:06:30 crc kubenswrapper[4934]: I1002 10:06:30.122883 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-62a6-account-create-nthhn"] Oct 02 10:06:30 crc kubenswrapper[4934]: I1002 10:06:30.123977 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-62a6-account-create-nthhn" Oct 02 10:06:30 crc kubenswrapper[4934]: I1002 10:06:30.128074 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 02 10:06:30 crc kubenswrapper[4934]: I1002 10:06:30.140680 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-62a6-account-create-nthhn"] Oct 02 10:06:30 crc kubenswrapper[4934]: I1002 10:06:30.152353 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nqjv\" (UniqueName: \"kubernetes.io/projected/3640bb49-1f9b-4bfd-9f11-4e6f5662bf76-kube-api-access-5nqjv\") pod \"barbican-62a6-account-create-nthhn\" (UID: \"3640bb49-1f9b-4bfd-9f11-4e6f5662bf76\") " pod="openstack/barbican-62a6-account-create-nthhn" Oct 02 10:06:30 crc kubenswrapper[4934]: I1002 10:06:30.253809 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nqjv\" (UniqueName: \"kubernetes.io/projected/3640bb49-1f9b-4bfd-9f11-4e6f5662bf76-kube-api-access-5nqjv\") pod \"barbican-62a6-account-create-nthhn\" (UID: \"3640bb49-1f9b-4bfd-9f11-4e6f5662bf76\") " pod="openstack/barbican-62a6-account-create-nthhn" Oct 02 10:06:30 crc kubenswrapper[4934]: I1002 10:06:30.259201 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-36fe-account-create-8rqjp" Oct 02 10:06:30 crc kubenswrapper[4934]: I1002 10:06:30.269180 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nqjv\" (UniqueName: \"kubernetes.io/projected/3640bb49-1f9b-4bfd-9f11-4e6f5662bf76-kube-api-access-5nqjv\") pod \"barbican-62a6-account-create-nthhn\" (UID: \"3640bb49-1f9b-4bfd-9f11-4e6f5662bf76\") " pod="openstack/barbican-62a6-account-create-nthhn" Oct 02 10:06:30 crc kubenswrapper[4934]: I1002 10:06:30.446165 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-62a6-account-create-nthhn" Oct 02 10:06:30 crc kubenswrapper[4934]: I1002 10:06:30.660815 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" Oct 02 10:06:30 crc kubenswrapper[4934]: I1002 10:06:30.718440 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-556b48f7c7-8rms5"] Oct 02 10:06:30 crc kubenswrapper[4934]: I1002 10:06:30.718731 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" podUID="19ef1848-83f1-4a21-903c-ea7d36522dd2" containerName="dnsmasq-dns" containerID="cri-o://5f23045960ea86318590a91a23756e660012b01bafb2ad070f7ad444363b5434" gracePeriod=10 Oct 02 10:06:33 crc kubenswrapper[4934]: I1002 10:06:33.867887 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" podUID="19ef1848-83f1-4a21-903c-ea7d36522dd2" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.116:5353: connect: connection refused" Oct 02 10:06:34 crc kubenswrapper[4934]: I1002 10:06:34.069379 4934 generic.go:334] "Generic (PLEG): container finished" podID="19ef1848-83f1-4a21-903c-ea7d36522dd2" containerID="5f23045960ea86318590a91a23756e660012b01bafb2ad070f7ad444363b5434" exitCode=0 Oct 02 10:06:34 crc kubenswrapper[4934]: I1002 10:06:34.069473 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" event={"ID":"19ef1848-83f1-4a21-903c-ea7d36522dd2","Type":"ContainerDied","Data":"5f23045960ea86318590a91a23756e660012b01bafb2ad070f7ad444363b5434"} Oct 02 10:06:35 crc kubenswrapper[4934]: I1002 10:06:35.003243 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-62a6-account-create-nthhn"] Oct 02 10:06:35 crc kubenswrapper[4934]: W1002 10:06:35.030977 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3640bb49_1f9b_4bfd_9f11_4e6f5662bf76.slice/crio-069a0bccb99c47020cc506fcf1694aa6b92a0c415c408344d715541ccc138cd7 WatchSource:0}: Error finding container 069a0bccb99c47020cc506fcf1694aa6b92a0c415c408344d715541ccc138cd7: Status 404 returned error can't find the container with id 069a0bccb99c47020cc506fcf1694aa6b92a0c415c408344d715541ccc138cd7 Oct 02 10:06:35 crc kubenswrapper[4934]: I1002 10:06:35.078487 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-62a6-account-create-nthhn" event={"ID":"3640bb49-1f9b-4bfd-9f11-4e6f5662bf76","Type":"ContainerStarted","Data":"069a0bccb99c47020cc506fcf1694aa6b92a0c415c408344d715541ccc138cd7"} Oct 02 10:06:35 crc kubenswrapper[4934]: I1002 10:06:35.252832 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-36fe-account-create-8rqjp"] Oct 02 10:06:35 crc kubenswrapper[4934]: W1002 10:06:35.255001 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe42c16a_682c_44b2_a53a_d33085d3a681.slice/crio-bffe9f19efefeb23b0490a68c4994dd8d85a2268769b8e2cd4f8a1c433d28172 WatchSource:0}: Error finding container bffe9f19efefeb23b0490a68c4994dd8d85a2268769b8e2cd4f8a1c433d28172: Status 404 returned error can't find the container with id bffe9f19efefeb23b0490a68c4994dd8d85a2268769b8e2cd4f8a1c433d28172 Oct 02 10:06:36 crc kubenswrapper[4934]: I1002 10:06:36.059600 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" Oct 02 10:06:36 crc kubenswrapper[4934]: I1002 10:06:36.095665 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-36fe-account-create-8rqjp" event={"ID":"be42c16a-682c-44b2-a53a-d33085d3a681","Type":"ContainerStarted","Data":"bffe9f19efefeb23b0490a68c4994dd8d85a2268769b8e2cd4f8a1c433d28172"} Oct 02 10:06:36 crc kubenswrapper[4934]: I1002 10:06:36.096976 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" event={"ID":"19ef1848-83f1-4a21-903c-ea7d36522dd2","Type":"ContainerDied","Data":"1661fc9453c0b8d374163b8cdc07bc1248e6c719538fc4e79f19967449e38ebf"} Oct 02 10:06:36 crc kubenswrapper[4934]: I1002 10:06:36.097015 4934 scope.go:117] "RemoveContainer" containerID="5f23045960ea86318590a91a23756e660012b01bafb2ad070f7ad444363b5434" Oct 02 10:06:36 crc kubenswrapper[4934]: I1002 10:06:36.097151 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-556b48f7c7-8rms5" Oct 02 10:06:36 crc kubenswrapper[4934]: I1002 10:06:36.147138 4934 scope.go:117] "RemoveContainer" containerID="076596e447df636fd2be21ec0842225ebdff1f702645ba2e8782db56398c415a" Oct 02 10:06:36 crc kubenswrapper[4934]: I1002 10:06:36.191495 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19ef1848-83f1-4a21-903c-ea7d36522dd2-dns-svc\") pod \"19ef1848-83f1-4a21-903c-ea7d36522dd2\" (UID: \"19ef1848-83f1-4a21-903c-ea7d36522dd2\") " Oct 02 10:06:36 crc kubenswrapper[4934]: I1002 10:06:36.191613 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/19ef1848-83f1-4a21-903c-ea7d36522dd2-ovsdbserver-sb\") pod \"19ef1848-83f1-4a21-903c-ea7d36522dd2\" (UID: \"19ef1848-83f1-4a21-903c-ea7d36522dd2\") " Oct 02 10:06:36 crc kubenswrapper[4934]: I1002 10:06:36.191685 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvzm6\" (UniqueName: \"kubernetes.io/projected/19ef1848-83f1-4a21-903c-ea7d36522dd2-kube-api-access-xvzm6\") pod \"19ef1848-83f1-4a21-903c-ea7d36522dd2\" (UID: \"19ef1848-83f1-4a21-903c-ea7d36522dd2\") " Oct 02 10:06:36 crc kubenswrapper[4934]: I1002 10:06:36.191748 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19ef1848-83f1-4a21-903c-ea7d36522dd2-config\") pod \"19ef1848-83f1-4a21-903c-ea7d36522dd2\" (UID: \"19ef1848-83f1-4a21-903c-ea7d36522dd2\") " Oct 02 10:06:36 crc kubenswrapper[4934]: I1002 10:06:36.191823 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19ef1848-83f1-4a21-903c-ea7d36522dd2-ovsdbserver-nb\") pod \"19ef1848-83f1-4a21-903c-ea7d36522dd2\" (UID: \"19ef1848-83f1-4a21-903c-ea7d36522dd2\") " Oct 02 10:06:36 crc kubenswrapper[4934]: I1002 10:06:36.197347 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19ef1848-83f1-4a21-903c-ea7d36522dd2-kube-api-access-xvzm6" (OuterVolumeSpecName: "kube-api-access-xvzm6") pod "19ef1848-83f1-4a21-903c-ea7d36522dd2" (UID: "19ef1848-83f1-4a21-903c-ea7d36522dd2"). InnerVolumeSpecName "kube-api-access-xvzm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:36 crc kubenswrapper[4934]: I1002 10:06:36.236362 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19ef1848-83f1-4a21-903c-ea7d36522dd2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "19ef1848-83f1-4a21-903c-ea7d36522dd2" (UID: "19ef1848-83f1-4a21-903c-ea7d36522dd2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:06:36 crc kubenswrapper[4934]: I1002 10:06:36.246627 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19ef1848-83f1-4a21-903c-ea7d36522dd2-config" (OuterVolumeSpecName: "config") pod "19ef1848-83f1-4a21-903c-ea7d36522dd2" (UID: "19ef1848-83f1-4a21-903c-ea7d36522dd2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:06:36 crc kubenswrapper[4934]: I1002 10:06:36.258200 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19ef1848-83f1-4a21-903c-ea7d36522dd2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "19ef1848-83f1-4a21-903c-ea7d36522dd2" (UID: "19ef1848-83f1-4a21-903c-ea7d36522dd2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:06:36 crc kubenswrapper[4934]: I1002 10:06:36.261400 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19ef1848-83f1-4a21-903c-ea7d36522dd2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "19ef1848-83f1-4a21-903c-ea7d36522dd2" (UID: "19ef1848-83f1-4a21-903c-ea7d36522dd2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:06:36 crc kubenswrapper[4934]: I1002 10:06:36.295230 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19ef1848-83f1-4a21-903c-ea7d36522dd2-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:36 crc kubenswrapper[4934]: I1002 10:06:36.295275 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/19ef1848-83f1-4a21-903c-ea7d36522dd2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:36 crc kubenswrapper[4934]: I1002 10:06:36.295300 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvzm6\" (UniqueName: \"kubernetes.io/projected/19ef1848-83f1-4a21-903c-ea7d36522dd2-kube-api-access-xvzm6\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:36 crc kubenswrapper[4934]: I1002 10:06:36.295318 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19ef1848-83f1-4a21-903c-ea7d36522dd2-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:36 crc kubenswrapper[4934]: I1002 10:06:36.295335 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19ef1848-83f1-4a21-903c-ea7d36522dd2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:36 crc kubenswrapper[4934]: I1002 10:06:36.443304 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-556b48f7c7-8rms5"] Oct 02 10:06:36 crc kubenswrapper[4934]: I1002 10:06:36.447059 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-556b48f7c7-8rms5"] Oct 02 10:06:36 crc kubenswrapper[4934]: E1002 10:06:36.773811 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api@sha256:85c8d86100270d60e99d2a7ab5e875fa7634a6a8a6c351630fe3b964e1b11f0e" Oct 02 10:06:36 crc kubenswrapper[4934]: E1002 10:06:36.774034 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:85c8d86100270d60e99d2a7ab5e875fa7634a6a8a6c351630fe3b964e1b11f0e,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zqq9t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-jwpkw_openstack(6427286d-a1ca-43ca-8ecb-bbdda45e360e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 10:06:36 crc kubenswrapper[4934]: E1002 10:06:36.775457 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-jwpkw" podUID="6427286d-a1ca-43ca-8ecb-bbdda45e360e" Oct 02 10:06:36 crc kubenswrapper[4934]: I1002 10:06:36.924883 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19ef1848-83f1-4a21-903c-ea7d36522dd2" path="/var/lib/kubelet/pods/19ef1848-83f1-4a21-903c-ea7d36522dd2/volumes" Oct 02 10:06:37 crc kubenswrapper[4934]: I1002 10:06:37.125654 4934 generic.go:334] "Generic (PLEG): container finished" podID="be42c16a-682c-44b2-a53a-d33085d3a681" containerID="df3044ab763e8645c1efadc31caa8506372fa484c120cc1970d19fc14e248262" exitCode=0 Oct 02 10:06:37 crc kubenswrapper[4934]: I1002 10:06:37.125844 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-36fe-account-create-8rqjp" event={"ID":"be42c16a-682c-44b2-a53a-d33085d3a681","Type":"ContainerDied","Data":"df3044ab763e8645c1efadc31caa8506372fa484c120cc1970d19fc14e248262"} Oct 02 10:06:37 crc kubenswrapper[4934]: I1002 10:06:37.129794 4934 generic.go:334] "Generic (PLEG): container finished" podID="3640bb49-1f9b-4bfd-9f11-4e6f5662bf76" containerID="b3acd6151738e0f4de858b51bf54757df9ef9b0a140e37a4f4a15c28e67237be" exitCode=0 Oct 02 10:06:37 crc kubenswrapper[4934]: I1002 10:06:37.129899 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-62a6-account-create-nthhn" event={"ID":"3640bb49-1f9b-4bfd-9f11-4e6f5662bf76","Type":"ContainerDied","Data":"b3acd6151738e0f4de858b51bf54757df9ef9b0a140e37a4f4a15c28e67237be"} Oct 02 10:06:37 crc kubenswrapper[4934]: E1002 10:06:37.133958 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api@sha256:85c8d86100270d60e99d2a7ab5e875fa7634a6a8a6c351630fe3b964e1b11f0e\\\"\"" pod="openstack/glance-db-sync-jwpkw" podUID="6427286d-a1ca-43ca-8ecb-bbdda45e360e" Oct 02 10:06:38 crc kubenswrapper[4934]: I1002 10:06:38.512148 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-36fe-account-create-8rqjp" Oct 02 10:06:38 crc kubenswrapper[4934]: I1002 10:06:38.518354 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-62a6-account-create-nthhn" Oct 02 10:06:38 crc kubenswrapper[4934]: I1002 10:06:38.633877 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nqjv\" (UniqueName: \"kubernetes.io/projected/3640bb49-1f9b-4bfd-9f11-4e6f5662bf76-kube-api-access-5nqjv\") pod \"3640bb49-1f9b-4bfd-9f11-4e6f5662bf76\" (UID: \"3640bb49-1f9b-4bfd-9f11-4e6f5662bf76\") " Oct 02 10:06:38 crc kubenswrapper[4934]: I1002 10:06:38.633966 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kdxj\" (UniqueName: \"kubernetes.io/projected/be42c16a-682c-44b2-a53a-d33085d3a681-kube-api-access-2kdxj\") pod \"be42c16a-682c-44b2-a53a-d33085d3a681\" (UID: \"be42c16a-682c-44b2-a53a-d33085d3a681\") " Oct 02 10:06:38 crc kubenswrapper[4934]: I1002 10:06:38.641464 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be42c16a-682c-44b2-a53a-d33085d3a681-kube-api-access-2kdxj" (OuterVolumeSpecName: "kube-api-access-2kdxj") pod "be42c16a-682c-44b2-a53a-d33085d3a681" (UID: "be42c16a-682c-44b2-a53a-d33085d3a681"). InnerVolumeSpecName "kube-api-access-2kdxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:38 crc kubenswrapper[4934]: I1002 10:06:38.642008 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3640bb49-1f9b-4bfd-9f11-4e6f5662bf76-kube-api-access-5nqjv" (OuterVolumeSpecName: "kube-api-access-5nqjv") pod "3640bb49-1f9b-4bfd-9f11-4e6f5662bf76" (UID: "3640bb49-1f9b-4bfd-9f11-4e6f5662bf76"). InnerVolumeSpecName "kube-api-access-5nqjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:38 crc kubenswrapper[4934]: I1002 10:06:38.736333 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nqjv\" (UniqueName: \"kubernetes.io/projected/3640bb49-1f9b-4bfd-9f11-4e6f5662bf76-kube-api-access-5nqjv\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:38 crc kubenswrapper[4934]: I1002 10:06:38.736381 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kdxj\" (UniqueName: \"kubernetes.io/projected/be42c16a-682c-44b2-a53a-d33085d3a681-kube-api-access-2kdxj\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:39 crc kubenswrapper[4934]: I1002 10:06:39.150456 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-62a6-account-create-nthhn" event={"ID":"3640bb49-1f9b-4bfd-9f11-4e6f5662bf76","Type":"ContainerDied","Data":"069a0bccb99c47020cc506fcf1694aa6b92a0c415c408344d715541ccc138cd7"} Oct 02 10:06:39 crc kubenswrapper[4934]: I1002 10:06:39.150508 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="069a0bccb99c47020cc506fcf1694aa6b92a0c415c408344d715541ccc138cd7" Oct 02 10:06:39 crc kubenswrapper[4934]: I1002 10:06:39.150482 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-62a6-account-create-nthhn" Oct 02 10:06:39 crc kubenswrapper[4934]: I1002 10:06:39.151972 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-36fe-account-create-8rqjp" event={"ID":"be42c16a-682c-44b2-a53a-d33085d3a681","Type":"ContainerDied","Data":"bffe9f19efefeb23b0490a68c4994dd8d85a2268769b8e2cd4f8a1c433d28172"} Oct 02 10:06:39 crc kubenswrapper[4934]: I1002 10:06:39.151991 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bffe9f19efefeb23b0490a68c4994dd8d85a2268769b8e2cd4f8a1c433d28172" Oct 02 10:06:39 crc kubenswrapper[4934]: I1002 10:06:39.152044 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-36fe-account-create-8rqjp" Oct 02 10:06:40 crc kubenswrapper[4934]: I1002 10:06:40.160913 4934 generic.go:334] "Generic (PLEG): container finished" podID="6f4da5aa-1b82-4286-8a3b-68ca8f486c78" containerID="8075fe4eb85efceb29e57e7b750c07c1739933c65104eec3beb30a995332f635" exitCode=0 Oct 02 10:06:40 crc kubenswrapper[4934]: I1002 10:06:40.161014 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-c25xv" event={"ID":"6f4da5aa-1b82-4286-8a3b-68ca8f486c78","Type":"ContainerDied","Data":"8075fe4eb85efceb29e57e7b750c07c1739933c65104eec3beb30a995332f635"} Oct 02 10:06:40 crc kubenswrapper[4934]: I1002 10:06:40.257888 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-70e2-account-create-cgqd8"] Oct 02 10:06:40 crc kubenswrapper[4934]: E1002 10:06:40.258299 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19ef1848-83f1-4a21-903c-ea7d36522dd2" containerName="dnsmasq-dns" Oct 02 10:06:40 crc kubenswrapper[4934]: I1002 10:06:40.258324 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="19ef1848-83f1-4a21-903c-ea7d36522dd2" containerName="dnsmasq-dns" Oct 02 10:06:40 crc kubenswrapper[4934]: E1002 10:06:40.258344 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be42c16a-682c-44b2-a53a-d33085d3a681" containerName="mariadb-account-create" Oct 02 10:06:40 crc kubenswrapper[4934]: I1002 10:06:40.258354 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="be42c16a-682c-44b2-a53a-d33085d3a681" containerName="mariadb-account-create" Oct 02 10:06:40 crc kubenswrapper[4934]: E1002 10:06:40.258374 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3640bb49-1f9b-4bfd-9f11-4e6f5662bf76" containerName="mariadb-account-create" Oct 02 10:06:40 crc kubenswrapper[4934]: I1002 10:06:40.258382 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="3640bb49-1f9b-4bfd-9f11-4e6f5662bf76" containerName="mariadb-account-create" Oct 02 10:06:40 crc kubenswrapper[4934]: E1002 10:06:40.258411 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19ef1848-83f1-4a21-903c-ea7d36522dd2" containerName="init" Oct 02 10:06:40 crc kubenswrapper[4934]: I1002 10:06:40.258419 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="19ef1848-83f1-4a21-903c-ea7d36522dd2" containerName="init" Oct 02 10:06:40 crc kubenswrapper[4934]: I1002 10:06:40.258614 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="be42c16a-682c-44b2-a53a-d33085d3a681" containerName="mariadb-account-create" Oct 02 10:06:40 crc kubenswrapper[4934]: I1002 10:06:40.258641 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="19ef1848-83f1-4a21-903c-ea7d36522dd2" containerName="dnsmasq-dns" Oct 02 10:06:40 crc kubenswrapper[4934]: I1002 10:06:40.258664 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="3640bb49-1f9b-4bfd-9f11-4e6f5662bf76" containerName="mariadb-account-create" Oct 02 10:06:40 crc kubenswrapper[4934]: I1002 10:06:40.259252 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-70e2-account-create-cgqd8" Oct 02 10:06:40 crc kubenswrapper[4934]: I1002 10:06:40.261628 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 02 10:06:40 crc kubenswrapper[4934]: I1002 10:06:40.268087 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-70e2-account-create-cgqd8"] Oct 02 10:06:40 crc kubenswrapper[4934]: I1002 10:06:40.362074 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2gn7\" (UniqueName: \"kubernetes.io/projected/f591dece-fc9a-430d-b0b4-5de71b477541-kube-api-access-w2gn7\") pod \"neutron-70e2-account-create-cgqd8\" (UID: \"f591dece-fc9a-430d-b0b4-5de71b477541\") " pod="openstack/neutron-70e2-account-create-cgqd8" Oct 02 10:06:40 crc kubenswrapper[4934]: I1002 10:06:40.463630 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2gn7\" (UniqueName: \"kubernetes.io/projected/f591dece-fc9a-430d-b0b4-5de71b477541-kube-api-access-w2gn7\") pod \"neutron-70e2-account-create-cgqd8\" (UID: \"f591dece-fc9a-430d-b0b4-5de71b477541\") " pod="openstack/neutron-70e2-account-create-cgqd8" Oct 02 10:06:40 crc kubenswrapper[4934]: I1002 10:06:40.479555 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2gn7\" (UniqueName: \"kubernetes.io/projected/f591dece-fc9a-430d-b0b4-5de71b477541-kube-api-access-w2gn7\") pod \"neutron-70e2-account-create-cgqd8\" (UID: \"f591dece-fc9a-430d-b0b4-5de71b477541\") " pod="openstack/neutron-70e2-account-create-cgqd8" Oct 02 10:06:40 crc kubenswrapper[4934]: I1002 10:06:40.580750 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-70e2-account-create-cgqd8" Oct 02 10:06:41 crc kubenswrapper[4934]: I1002 10:06:41.016443 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-70e2-account-create-cgqd8"] Oct 02 10:06:41 crc kubenswrapper[4934]: I1002 10:06:41.170441 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-70e2-account-create-cgqd8" event={"ID":"f591dece-fc9a-430d-b0b4-5de71b477541","Type":"ContainerStarted","Data":"b3eafddce2353e95475b8ad7ef49b05c5e4af10b9952dd93b698aecd957511bb"} Oct 02 10:06:41 crc kubenswrapper[4934]: I1002 10:06:41.386858 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-c25xv" Oct 02 10:06:41 crc kubenswrapper[4934]: I1002 10:06:41.478309 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f4da5aa-1b82-4286-8a3b-68ca8f486c78-combined-ca-bundle\") pod \"6f4da5aa-1b82-4286-8a3b-68ca8f486c78\" (UID: \"6f4da5aa-1b82-4286-8a3b-68ca8f486c78\") " Oct 02 10:06:41 crc kubenswrapper[4934]: I1002 10:06:41.478480 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f4da5aa-1b82-4286-8a3b-68ca8f486c78-config-data\") pod \"6f4da5aa-1b82-4286-8a3b-68ca8f486c78\" (UID: \"6f4da5aa-1b82-4286-8a3b-68ca8f486c78\") " Oct 02 10:06:41 crc kubenswrapper[4934]: I1002 10:06:41.478528 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66gz8\" (UniqueName: \"kubernetes.io/projected/6f4da5aa-1b82-4286-8a3b-68ca8f486c78-kube-api-access-66gz8\") pod \"6f4da5aa-1b82-4286-8a3b-68ca8f486c78\" (UID: \"6f4da5aa-1b82-4286-8a3b-68ca8f486c78\") " Oct 02 10:06:41 crc kubenswrapper[4934]: I1002 10:06:41.485865 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f4da5aa-1b82-4286-8a3b-68ca8f486c78-kube-api-access-66gz8" (OuterVolumeSpecName: "kube-api-access-66gz8") pod "6f4da5aa-1b82-4286-8a3b-68ca8f486c78" (UID: "6f4da5aa-1b82-4286-8a3b-68ca8f486c78"). InnerVolumeSpecName "kube-api-access-66gz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:41 crc kubenswrapper[4934]: I1002 10:06:41.509322 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f4da5aa-1b82-4286-8a3b-68ca8f486c78-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6f4da5aa-1b82-4286-8a3b-68ca8f486c78" (UID: "6f4da5aa-1b82-4286-8a3b-68ca8f486c78"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:41 crc kubenswrapper[4934]: I1002 10:06:41.530383 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f4da5aa-1b82-4286-8a3b-68ca8f486c78-config-data" (OuterVolumeSpecName: "config-data") pod "6f4da5aa-1b82-4286-8a3b-68ca8f486c78" (UID: "6f4da5aa-1b82-4286-8a3b-68ca8f486c78"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:41 crc kubenswrapper[4934]: I1002 10:06:41.580076 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f4da5aa-1b82-4286-8a3b-68ca8f486c78-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:41 crc kubenswrapper[4934]: I1002 10:06:41.580106 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f4da5aa-1b82-4286-8a3b-68ca8f486c78-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:41 crc kubenswrapper[4934]: I1002 10:06:41.580115 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66gz8\" (UniqueName: \"kubernetes.io/projected/6f4da5aa-1b82-4286-8a3b-68ca8f486c78-kube-api-access-66gz8\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.179825 4934 generic.go:334] "Generic (PLEG): container finished" podID="f591dece-fc9a-430d-b0b4-5de71b477541" containerID="d6e70f3f0fffee829ba104475e221e24a859707e848e6edd6c64293f59aa17d9" exitCode=0 Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.179866 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-70e2-account-create-cgqd8" event={"ID":"f591dece-fc9a-430d-b0b4-5de71b477541","Type":"ContainerDied","Data":"d6e70f3f0fffee829ba104475e221e24a859707e848e6edd6c64293f59aa17d9"} Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.181846 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-c25xv" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.181869 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-c25xv" event={"ID":"6f4da5aa-1b82-4286-8a3b-68ca8f486c78","Type":"ContainerDied","Data":"8f3db57be78c368ad20e5871f54d698c79fe3ee4af0f516f759204c87c225a64"} Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.181896 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f3db57be78c368ad20e5871f54d698c79fe3ee4af0f516f759204c87c225a64" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.408052 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f6d4499cf-btbf8"] Oct 02 10:06:42 crc kubenswrapper[4934]: E1002 10:06:42.410130 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f4da5aa-1b82-4286-8a3b-68ca8f486c78" containerName="keystone-db-sync" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.410164 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f4da5aa-1b82-4286-8a3b-68ca8f486c78" containerName="keystone-db-sync" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.410423 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f4da5aa-1b82-4286-8a3b-68ca8f486c78" containerName="keystone-db-sync" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.415121 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f6d4499cf-btbf8" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.422781 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f6d4499cf-btbf8"] Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.491192 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-xbljs"] Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.492491 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xbljs" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.502467 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-xbljs"] Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.503447 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.503739 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.504012 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.504375 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-bkm4l" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.597369 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rp9rg\" (UniqueName: \"kubernetes.io/projected/6f91778c-ae60-410b-8a84-16007820b80b-kube-api-access-rp9rg\") pod \"keystone-bootstrap-xbljs\" (UID: \"6f91778c-ae60-410b-8a84-16007820b80b\") " pod="openstack/keystone-bootstrap-xbljs" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.597429 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-ovsdbserver-nb\") pod \"dnsmasq-dns-f6d4499cf-btbf8\" (UID: \"be49fd0e-9182-46e3-aa25-4b2681694826\") " pod="openstack/dnsmasq-dns-f6d4499cf-btbf8" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.597466 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-combined-ca-bundle\") pod \"keystone-bootstrap-xbljs\" (UID: \"6f91778c-ae60-410b-8a84-16007820b80b\") " pod="openstack/keystone-bootstrap-xbljs" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.597519 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-dns-svc\") pod \"dnsmasq-dns-f6d4499cf-btbf8\" (UID: \"be49fd0e-9182-46e3-aa25-4b2681694826\") " pod="openstack/dnsmasq-dns-f6d4499cf-btbf8" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.597683 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-config-data\") pod \"keystone-bootstrap-xbljs\" (UID: \"6f91778c-ae60-410b-8a84-16007820b80b\") " pod="openstack/keystone-bootstrap-xbljs" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.597726 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-credential-keys\") pod \"keystone-bootstrap-xbljs\" (UID: \"6f91778c-ae60-410b-8a84-16007820b80b\") " pod="openstack/keystone-bootstrap-xbljs" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.597902 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-ovsdbserver-sb\") pod \"dnsmasq-dns-f6d4499cf-btbf8\" (UID: \"be49fd0e-9182-46e3-aa25-4b2681694826\") " pod="openstack/dnsmasq-dns-f6d4499cf-btbf8" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.598020 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-scripts\") pod \"keystone-bootstrap-xbljs\" (UID: \"6f91778c-ae60-410b-8a84-16007820b80b\") " pod="openstack/keystone-bootstrap-xbljs" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.598068 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-fernet-keys\") pod \"keystone-bootstrap-xbljs\" (UID: \"6f91778c-ae60-410b-8a84-16007820b80b\") " pod="openstack/keystone-bootstrap-xbljs" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.598163 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-dns-swift-storage-0\") pod \"dnsmasq-dns-f6d4499cf-btbf8\" (UID: \"be49fd0e-9182-46e3-aa25-4b2681694826\") " pod="openstack/dnsmasq-dns-f6d4499cf-btbf8" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.598230 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6tjx\" (UniqueName: \"kubernetes.io/projected/be49fd0e-9182-46e3-aa25-4b2681694826-kube-api-access-v6tjx\") pod \"dnsmasq-dns-f6d4499cf-btbf8\" (UID: \"be49fd0e-9182-46e3-aa25-4b2681694826\") " pod="openstack/dnsmasq-dns-f6d4499cf-btbf8" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.598268 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-config\") pod \"dnsmasq-dns-f6d4499cf-btbf8\" (UID: \"be49fd0e-9182-46e3-aa25-4b2681694826\") " pod="openstack/dnsmasq-dns-f6d4499cf-btbf8" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.639331 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.641357 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.647223 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.647380 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.652318 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-kb5lq"] Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.653697 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-kb5lq" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.663088 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.663267 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.663367 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-b6f84" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.669647 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.682243 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-kb5lq"] Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.699890 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6tjx\" (UniqueName: \"kubernetes.io/projected/be49fd0e-9182-46e3-aa25-4b2681694826-kube-api-access-v6tjx\") pod \"dnsmasq-dns-f6d4499cf-btbf8\" (UID: \"be49fd0e-9182-46e3-aa25-4b2681694826\") " pod="openstack/dnsmasq-dns-f6d4499cf-btbf8" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.699945 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-config\") pod \"dnsmasq-dns-f6d4499cf-btbf8\" (UID: \"be49fd0e-9182-46e3-aa25-4b2681694826\") " pod="openstack/dnsmasq-dns-f6d4499cf-btbf8" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.699974 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rp9rg\" (UniqueName: \"kubernetes.io/projected/6f91778c-ae60-410b-8a84-16007820b80b-kube-api-access-rp9rg\") pod \"keystone-bootstrap-xbljs\" (UID: \"6f91778c-ae60-410b-8a84-16007820b80b\") " pod="openstack/keystone-bootstrap-xbljs" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.699995 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-ovsdbserver-nb\") pod \"dnsmasq-dns-f6d4499cf-btbf8\" (UID: \"be49fd0e-9182-46e3-aa25-4b2681694826\") " pod="openstack/dnsmasq-dns-f6d4499cf-btbf8" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.700012 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-combined-ca-bundle\") pod \"keystone-bootstrap-xbljs\" (UID: \"6f91778c-ae60-410b-8a84-16007820b80b\") " pod="openstack/keystone-bootstrap-xbljs" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.700034 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-dns-svc\") pod \"dnsmasq-dns-f6d4499cf-btbf8\" (UID: \"be49fd0e-9182-46e3-aa25-4b2681694826\") " pod="openstack/dnsmasq-dns-f6d4499cf-btbf8" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.700064 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-config-data\") pod \"keystone-bootstrap-xbljs\" (UID: \"6f91778c-ae60-410b-8a84-16007820b80b\") " pod="openstack/keystone-bootstrap-xbljs" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.700078 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-credential-keys\") pod \"keystone-bootstrap-xbljs\" (UID: \"6f91778c-ae60-410b-8a84-16007820b80b\") " pod="openstack/keystone-bootstrap-xbljs" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.700107 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-ovsdbserver-sb\") pod \"dnsmasq-dns-f6d4499cf-btbf8\" (UID: \"be49fd0e-9182-46e3-aa25-4b2681694826\") " pod="openstack/dnsmasq-dns-f6d4499cf-btbf8" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.700140 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-scripts\") pod \"keystone-bootstrap-xbljs\" (UID: \"6f91778c-ae60-410b-8a84-16007820b80b\") " pod="openstack/keystone-bootstrap-xbljs" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.700157 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-fernet-keys\") pod \"keystone-bootstrap-xbljs\" (UID: \"6f91778c-ae60-410b-8a84-16007820b80b\") " pod="openstack/keystone-bootstrap-xbljs" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.700188 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-dns-swift-storage-0\") pod \"dnsmasq-dns-f6d4499cf-btbf8\" (UID: \"be49fd0e-9182-46e3-aa25-4b2681694826\") " pod="openstack/dnsmasq-dns-f6d4499cf-btbf8" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.700993 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-dns-swift-storage-0\") pod \"dnsmasq-dns-f6d4499cf-btbf8\" (UID: \"be49fd0e-9182-46e3-aa25-4b2681694826\") " pod="openstack/dnsmasq-dns-f6d4499cf-btbf8" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.701797 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-config\") pod \"dnsmasq-dns-f6d4499cf-btbf8\" (UID: \"be49fd0e-9182-46e3-aa25-4b2681694826\") " pod="openstack/dnsmasq-dns-f6d4499cf-btbf8" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.704408 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-dns-svc\") pod \"dnsmasq-dns-f6d4499cf-btbf8\" (UID: \"be49fd0e-9182-46e3-aa25-4b2681694826\") " pod="openstack/dnsmasq-dns-f6d4499cf-btbf8" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.707252 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-ovsdbserver-nb\") pod \"dnsmasq-dns-f6d4499cf-btbf8\" (UID: \"be49fd0e-9182-46e3-aa25-4b2681694826\") " pod="openstack/dnsmasq-dns-f6d4499cf-btbf8" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.708284 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-ovsdbserver-sb\") pod \"dnsmasq-dns-f6d4499cf-btbf8\" (UID: \"be49fd0e-9182-46e3-aa25-4b2681694826\") " pod="openstack/dnsmasq-dns-f6d4499cf-btbf8" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.721145 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-config-data\") pod \"keystone-bootstrap-xbljs\" (UID: \"6f91778c-ae60-410b-8a84-16007820b80b\") " pod="openstack/keystone-bootstrap-xbljs" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.722073 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-credential-keys\") pod \"keystone-bootstrap-xbljs\" (UID: \"6f91778c-ae60-410b-8a84-16007820b80b\") " pod="openstack/keystone-bootstrap-xbljs" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.725175 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-fernet-keys\") pod \"keystone-bootstrap-xbljs\" (UID: \"6f91778c-ae60-410b-8a84-16007820b80b\") " pod="openstack/keystone-bootstrap-xbljs" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.732447 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6tjx\" (UniqueName: \"kubernetes.io/projected/be49fd0e-9182-46e3-aa25-4b2681694826-kube-api-access-v6tjx\") pod \"dnsmasq-dns-f6d4499cf-btbf8\" (UID: \"be49fd0e-9182-46e3-aa25-4b2681694826\") " pod="openstack/dnsmasq-dns-f6d4499cf-btbf8" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.732871 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f6d4499cf-btbf8" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.735475 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-combined-ca-bundle\") pod \"keystone-bootstrap-xbljs\" (UID: \"6f91778c-ae60-410b-8a84-16007820b80b\") " pod="openstack/keystone-bootstrap-xbljs" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.739742 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-scripts\") pod \"keystone-bootstrap-xbljs\" (UID: \"6f91778c-ae60-410b-8a84-16007820b80b\") " pod="openstack/keystone-bootstrap-xbljs" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.739994 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-rqm5f"] Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.740441 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rp9rg\" (UniqueName: \"kubernetes.io/projected/6f91778c-ae60-410b-8a84-16007820b80b-kube-api-access-rp9rg\") pod \"keystone-bootstrap-xbljs\" (UID: \"6f91778c-ae60-410b-8a84-16007820b80b\") " pod="openstack/keystone-bootstrap-xbljs" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.741341 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-rqm5f" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.751009 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-rqm5f"] Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.753364 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-5l4m9" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.754029 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.802558 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-scripts\") pod \"cinder-db-sync-kb5lq\" (UID: \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\") " pod="openstack/cinder-db-sync-kb5lq" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.802679 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlskg\" (UniqueName: \"kubernetes.io/projected/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-kube-api-access-qlskg\") pod \"cinder-db-sync-kb5lq\" (UID: \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\") " pod="openstack/cinder-db-sync-kb5lq" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.802718 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a01b1d99-451c-4aec-b261-53232f25e582-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a01b1d99-451c-4aec-b261-53232f25e582\") " pod="openstack/ceilometer-0" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.802801 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-combined-ca-bundle\") pod \"cinder-db-sync-kb5lq\" (UID: \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\") " pod="openstack/cinder-db-sync-kb5lq" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.802841 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-config-data\") pod \"cinder-db-sync-kb5lq\" (UID: \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\") " pod="openstack/cinder-db-sync-kb5lq" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.802892 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a01b1d99-451c-4aec-b261-53232f25e582-scripts\") pod \"ceilometer-0\" (UID: \"a01b1d99-451c-4aec-b261-53232f25e582\") " pod="openstack/ceilometer-0" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.802928 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a01b1d99-451c-4aec-b261-53232f25e582-log-httpd\") pod \"ceilometer-0\" (UID: \"a01b1d99-451c-4aec-b261-53232f25e582\") " pod="openstack/ceilometer-0" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.802962 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qtm6\" (UniqueName: \"kubernetes.io/projected/a01b1d99-451c-4aec-b261-53232f25e582-kube-api-access-5qtm6\") pod \"ceilometer-0\" (UID: \"a01b1d99-451c-4aec-b261-53232f25e582\") " pod="openstack/ceilometer-0" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.802994 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-etc-machine-id\") pod \"cinder-db-sync-kb5lq\" (UID: \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\") " pod="openstack/cinder-db-sync-kb5lq" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.803022 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a01b1d99-451c-4aec-b261-53232f25e582-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a01b1d99-451c-4aec-b261-53232f25e582\") " pod="openstack/ceilometer-0" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.803098 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a01b1d99-451c-4aec-b261-53232f25e582-run-httpd\") pod \"ceilometer-0\" (UID: \"a01b1d99-451c-4aec-b261-53232f25e582\") " pod="openstack/ceilometer-0" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.803114 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-db-sync-config-data\") pod \"cinder-db-sync-kb5lq\" (UID: \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\") " pod="openstack/cinder-db-sync-kb5lq" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.803139 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a01b1d99-451c-4aec-b261-53232f25e582-config-data\") pod \"ceilometer-0\" (UID: \"a01b1d99-451c-4aec-b261-53232f25e582\") " pod="openstack/ceilometer-0" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.810107 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xbljs" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.843308 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f6d4499cf-btbf8"] Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.891638 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-685744b6b9-5mdh2"] Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.895867 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.909067 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-etc-machine-id\") pod \"cinder-db-sync-kb5lq\" (UID: \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\") " pod="openstack/cinder-db-sync-kb5lq" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.909132 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a01b1d99-451c-4aec-b261-53232f25e582-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a01b1d99-451c-4aec-b261-53232f25e582\") " pod="openstack/ceilometer-0" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.909167 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a01b1d99-451c-4aec-b261-53232f25e582-run-httpd\") pod \"ceilometer-0\" (UID: \"a01b1d99-451c-4aec-b261-53232f25e582\") " pod="openstack/ceilometer-0" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.909214 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-db-sync-config-data\") pod \"cinder-db-sync-kb5lq\" (UID: \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\") " pod="openstack/cinder-db-sync-kb5lq" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.909255 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a01b1d99-451c-4aec-b261-53232f25e582-config-data\") pod \"ceilometer-0\" (UID: \"a01b1d99-451c-4aec-b261-53232f25e582\") " pod="openstack/ceilometer-0" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.909284 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmzvx\" (UniqueName: \"kubernetes.io/projected/379bc476-7f65-408d-929a-8badb1e4dfdc-kube-api-access-bmzvx\") pod \"barbican-db-sync-rqm5f\" (UID: \"379bc476-7f65-408d-929a-8badb1e4dfdc\") " pod="openstack/barbican-db-sync-rqm5f" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.909339 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-scripts\") pod \"cinder-db-sync-kb5lq\" (UID: \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\") " pod="openstack/cinder-db-sync-kb5lq" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.909369 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/379bc476-7f65-408d-929a-8badb1e4dfdc-combined-ca-bundle\") pod \"barbican-db-sync-rqm5f\" (UID: \"379bc476-7f65-408d-929a-8badb1e4dfdc\") " pod="openstack/barbican-db-sync-rqm5f" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.909427 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlskg\" (UniqueName: \"kubernetes.io/projected/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-kube-api-access-qlskg\") pod \"cinder-db-sync-kb5lq\" (UID: \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\") " pod="openstack/cinder-db-sync-kb5lq" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.909458 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a01b1d99-451c-4aec-b261-53232f25e582-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a01b1d99-451c-4aec-b261-53232f25e582\") " pod="openstack/ceilometer-0" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.909511 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-combined-ca-bundle\") pod \"cinder-db-sync-kb5lq\" (UID: \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\") " pod="openstack/cinder-db-sync-kb5lq" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.909539 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-config-data\") pod \"cinder-db-sync-kb5lq\" (UID: \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\") " pod="openstack/cinder-db-sync-kb5lq" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.909569 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/379bc476-7f65-408d-929a-8badb1e4dfdc-db-sync-config-data\") pod \"barbican-db-sync-rqm5f\" (UID: \"379bc476-7f65-408d-929a-8badb1e4dfdc\") " pod="openstack/barbican-db-sync-rqm5f" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.909708 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a01b1d99-451c-4aec-b261-53232f25e582-scripts\") pod \"ceilometer-0\" (UID: \"a01b1d99-451c-4aec-b261-53232f25e582\") " pod="openstack/ceilometer-0" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.909738 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a01b1d99-451c-4aec-b261-53232f25e582-log-httpd\") pod \"ceilometer-0\" (UID: \"a01b1d99-451c-4aec-b261-53232f25e582\") " pod="openstack/ceilometer-0" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.909772 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qtm6\" (UniqueName: \"kubernetes.io/projected/a01b1d99-451c-4aec-b261-53232f25e582-kube-api-access-5qtm6\") pod \"ceilometer-0\" (UID: \"a01b1d99-451c-4aec-b261-53232f25e582\") " pod="openstack/ceilometer-0" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.910698 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-klshd"] Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.911108 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a01b1d99-451c-4aec-b261-53232f25e582-run-httpd\") pod \"ceilometer-0\" (UID: \"a01b1d99-451c-4aec-b261-53232f25e582\") " pod="openstack/ceilometer-0" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.911173 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-etc-machine-id\") pod \"cinder-db-sync-kb5lq\" (UID: \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\") " pod="openstack/cinder-db-sync-kb5lq" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.912106 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a01b1d99-451c-4aec-b261-53232f25e582-log-httpd\") pod \"ceilometer-0\" (UID: \"a01b1d99-451c-4aec-b261-53232f25e582\") " pod="openstack/ceilometer-0" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.920092 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-klshd" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.921408 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-combined-ca-bundle\") pod \"cinder-db-sync-kb5lq\" (UID: \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\") " pod="openstack/cinder-db-sync-kb5lq" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.926921 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.927044 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a01b1d99-451c-4aec-b261-53232f25e582-scripts\") pod \"ceilometer-0\" (UID: \"a01b1d99-451c-4aec-b261-53232f25e582\") " pod="openstack/ceilometer-0" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.927112 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-t9vdb" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.927407 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-config-data\") pod \"cinder-db-sync-kb5lq\" (UID: \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\") " pod="openstack/cinder-db-sync-kb5lq" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.927184 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.933481 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a01b1d99-451c-4aec-b261-53232f25e582-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a01b1d99-451c-4aec-b261-53232f25e582\") " pod="openstack/ceilometer-0" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.976504 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a01b1d99-451c-4aec-b261-53232f25e582-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a01b1d99-451c-4aec-b261-53232f25e582\") " pod="openstack/ceilometer-0" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.978285 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a01b1d99-451c-4aec-b261-53232f25e582-config-data\") pod \"ceilometer-0\" (UID: \"a01b1d99-451c-4aec-b261-53232f25e582\") " pod="openstack/ceilometer-0" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.979615 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-db-sync-config-data\") pod \"cinder-db-sync-kb5lq\" (UID: \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\") " pod="openstack/cinder-db-sync-kb5lq" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.980430 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qtm6\" (UniqueName: \"kubernetes.io/projected/a01b1d99-451c-4aec-b261-53232f25e582-kube-api-access-5qtm6\") pod \"ceilometer-0\" (UID: \"a01b1d99-451c-4aec-b261-53232f25e582\") " pod="openstack/ceilometer-0" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.983084 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlskg\" (UniqueName: \"kubernetes.io/projected/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-kube-api-access-qlskg\") pod \"cinder-db-sync-kb5lq\" (UID: \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\") " pod="openstack/cinder-db-sync-kb5lq" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.984755 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-klshd"] Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.985516 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.986900 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-scripts\") pod \"cinder-db-sync-kb5lq\" (UID: \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\") " pod="openstack/cinder-db-sync-kb5lq" Oct 02 10:06:42 crc kubenswrapper[4934]: I1002 10:06:42.991332 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-685744b6b9-5mdh2"] Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.012299 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-dns-swift-storage-0\") pod \"dnsmasq-dns-685744b6b9-5mdh2\" (UID: \"f2059861-459c-448a-933d-bc0d7660a9dc\") " pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.012364 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-ovsdbserver-nb\") pod \"dnsmasq-dns-685744b6b9-5mdh2\" (UID: \"f2059861-459c-448a-933d-bc0d7660a9dc\") " pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.012411 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmzvx\" (UniqueName: \"kubernetes.io/projected/379bc476-7f65-408d-929a-8badb1e4dfdc-kube-api-access-bmzvx\") pod \"barbican-db-sync-rqm5f\" (UID: \"379bc476-7f65-408d-929a-8badb1e4dfdc\") " pod="openstack/barbican-db-sync-rqm5f" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.012497 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/157aa366-544e-42ba-bd81-1649575c976f-combined-ca-bundle\") pod \"placement-db-sync-klshd\" (UID: \"157aa366-544e-42ba-bd81-1649575c976f\") " pod="openstack/placement-db-sync-klshd" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.012540 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-ovsdbserver-sb\") pod \"dnsmasq-dns-685744b6b9-5mdh2\" (UID: \"f2059861-459c-448a-933d-bc0d7660a9dc\") " pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.012629 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/379bc476-7f65-408d-929a-8badb1e4dfdc-combined-ca-bundle\") pod \"barbican-db-sync-rqm5f\" (UID: \"379bc476-7f65-408d-929a-8badb1e4dfdc\") " pod="openstack/barbican-db-sync-rqm5f" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.012677 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/157aa366-544e-42ba-bd81-1649575c976f-config-data\") pod \"placement-db-sync-klshd\" (UID: \"157aa366-544e-42ba-bd81-1649575c976f\") " pod="openstack/placement-db-sync-klshd" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.012869 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/157aa366-544e-42ba-bd81-1649575c976f-logs\") pod \"placement-db-sync-klshd\" (UID: \"157aa366-544e-42ba-bd81-1649575c976f\") " pod="openstack/placement-db-sync-klshd" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.012925 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-dns-svc\") pod \"dnsmasq-dns-685744b6b9-5mdh2\" (UID: \"f2059861-459c-448a-933d-bc0d7660a9dc\") " pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.012976 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jp9h\" (UniqueName: \"kubernetes.io/projected/157aa366-544e-42ba-bd81-1649575c976f-kube-api-access-4jp9h\") pod \"placement-db-sync-klshd\" (UID: \"157aa366-544e-42ba-bd81-1649575c976f\") " pod="openstack/placement-db-sync-klshd" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.013045 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/157aa366-544e-42ba-bd81-1649575c976f-scripts\") pod \"placement-db-sync-klshd\" (UID: \"157aa366-544e-42ba-bd81-1649575c976f\") " pod="openstack/placement-db-sync-klshd" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.013076 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-config\") pod \"dnsmasq-dns-685744b6b9-5mdh2\" (UID: \"f2059861-459c-448a-933d-bc0d7660a9dc\") " pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.013115 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/379bc476-7f65-408d-929a-8badb1e4dfdc-db-sync-config-data\") pod \"barbican-db-sync-rqm5f\" (UID: \"379bc476-7f65-408d-929a-8badb1e4dfdc\") " pod="openstack/barbican-db-sync-rqm5f" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.013186 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbxrb\" (UniqueName: \"kubernetes.io/projected/f2059861-459c-448a-933d-bc0d7660a9dc-kube-api-access-fbxrb\") pod \"dnsmasq-dns-685744b6b9-5mdh2\" (UID: \"f2059861-459c-448a-933d-bc0d7660a9dc\") " pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.019223 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/379bc476-7f65-408d-929a-8badb1e4dfdc-db-sync-config-data\") pod \"barbican-db-sync-rqm5f\" (UID: \"379bc476-7f65-408d-929a-8badb1e4dfdc\") " pod="openstack/barbican-db-sync-rqm5f" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.019276 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/379bc476-7f65-408d-929a-8badb1e4dfdc-combined-ca-bundle\") pod \"barbican-db-sync-rqm5f\" (UID: \"379bc476-7f65-408d-929a-8badb1e4dfdc\") " pod="openstack/barbican-db-sync-rqm5f" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.031747 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmzvx\" (UniqueName: \"kubernetes.io/projected/379bc476-7f65-408d-929a-8badb1e4dfdc-kube-api-access-bmzvx\") pod \"barbican-db-sync-rqm5f\" (UID: \"379bc476-7f65-408d-929a-8badb1e4dfdc\") " pod="openstack/barbican-db-sync-rqm5f" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.114638 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-dns-swift-storage-0\") pod \"dnsmasq-dns-685744b6b9-5mdh2\" (UID: \"f2059861-459c-448a-933d-bc0d7660a9dc\") " pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.114883 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-ovsdbserver-nb\") pod \"dnsmasq-dns-685744b6b9-5mdh2\" (UID: \"f2059861-459c-448a-933d-bc0d7660a9dc\") " pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.114918 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/157aa366-544e-42ba-bd81-1649575c976f-combined-ca-bundle\") pod \"placement-db-sync-klshd\" (UID: \"157aa366-544e-42ba-bd81-1649575c976f\") " pod="openstack/placement-db-sync-klshd" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.114933 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-ovsdbserver-sb\") pod \"dnsmasq-dns-685744b6b9-5mdh2\" (UID: \"f2059861-459c-448a-933d-bc0d7660a9dc\") " pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.114957 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/157aa366-544e-42ba-bd81-1649575c976f-config-data\") pod \"placement-db-sync-klshd\" (UID: \"157aa366-544e-42ba-bd81-1649575c976f\") " pod="openstack/placement-db-sync-klshd" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.114975 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/157aa366-544e-42ba-bd81-1649575c976f-logs\") pod \"placement-db-sync-klshd\" (UID: \"157aa366-544e-42ba-bd81-1649575c976f\") " pod="openstack/placement-db-sync-klshd" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.114995 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-dns-svc\") pod \"dnsmasq-dns-685744b6b9-5mdh2\" (UID: \"f2059861-459c-448a-933d-bc0d7660a9dc\") " pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.115048 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jp9h\" (UniqueName: \"kubernetes.io/projected/157aa366-544e-42ba-bd81-1649575c976f-kube-api-access-4jp9h\") pod \"placement-db-sync-klshd\" (UID: \"157aa366-544e-42ba-bd81-1649575c976f\") " pod="openstack/placement-db-sync-klshd" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.115090 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/157aa366-544e-42ba-bd81-1649575c976f-scripts\") pod \"placement-db-sync-klshd\" (UID: \"157aa366-544e-42ba-bd81-1649575c976f\") " pod="openstack/placement-db-sync-klshd" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.115110 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-config\") pod \"dnsmasq-dns-685744b6b9-5mdh2\" (UID: \"f2059861-459c-448a-933d-bc0d7660a9dc\") " pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.115152 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbxrb\" (UniqueName: \"kubernetes.io/projected/f2059861-459c-448a-933d-bc0d7660a9dc-kube-api-access-fbxrb\") pod \"dnsmasq-dns-685744b6b9-5mdh2\" (UID: \"f2059861-459c-448a-933d-bc0d7660a9dc\") " pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.115986 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-ovsdbserver-sb\") pod \"dnsmasq-dns-685744b6b9-5mdh2\" (UID: \"f2059861-459c-448a-933d-bc0d7660a9dc\") " pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.116795 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-ovsdbserver-nb\") pod \"dnsmasq-dns-685744b6b9-5mdh2\" (UID: \"f2059861-459c-448a-933d-bc0d7660a9dc\") " pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.116841 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/157aa366-544e-42ba-bd81-1649575c976f-logs\") pod \"placement-db-sync-klshd\" (UID: \"157aa366-544e-42ba-bd81-1649575c976f\") " pod="openstack/placement-db-sync-klshd" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.118514 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-dns-svc\") pod \"dnsmasq-dns-685744b6b9-5mdh2\" (UID: \"f2059861-459c-448a-933d-bc0d7660a9dc\") " pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.119867 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-config\") pod \"dnsmasq-dns-685744b6b9-5mdh2\" (UID: \"f2059861-459c-448a-933d-bc0d7660a9dc\") " pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.120103 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-dns-swift-storage-0\") pod \"dnsmasq-dns-685744b6b9-5mdh2\" (UID: \"f2059861-459c-448a-933d-bc0d7660a9dc\") " pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.123968 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/157aa366-544e-42ba-bd81-1649575c976f-scripts\") pod \"placement-db-sync-klshd\" (UID: \"157aa366-544e-42ba-bd81-1649575c976f\") " pod="openstack/placement-db-sync-klshd" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.124020 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/157aa366-544e-42ba-bd81-1649575c976f-config-data\") pod \"placement-db-sync-klshd\" (UID: \"157aa366-544e-42ba-bd81-1649575c976f\") " pod="openstack/placement-db-sync-klshd" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.124871 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/157aa366-544e-42ba-bd81-1649575c976f-combined-ca-bundle\") pod \"placement-db-sync-klshd\" (UID: \"157aa366-544e-42ba-bd81-1649575c976f\") " pod="openstack/placement-db-sync-klshd" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.135352 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jp9h\" (UniqueName: \"kubernetes.io/projected/157aa366-544e-42ba-bd81-1649575c976f-kube-api-access-4jp9h\") pod \"placement-db-sync-klshd\" (UID: \"157aa366-544e-42ba-bd81-1649575c976f\") " pod="openstack/placement-db-sync-klshd" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.139334 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbxrb\" (UniqueName: \"kubernetes.io/projected/f2059861-459c-448a-933d-bc0d7660a9dc-kube-api-access-fbxrb\") pod \"dnsmasq-dns-685744b6b9-5mdh2\" (UID: \"f2059861-459c-448a-933d-bc0d7660a9dc\") " pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.212426 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-rqm5f" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.278208 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-kb5lq" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.322723 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f6d4499cf-btbf8"] Oct 02 10:06:43 crc kubenswrapper[4934]: W1002 10:06:43.334364 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe49fd0e_9182_46e3_aa25_4b2681694826.slice/crio-a867f7b960475fce98aca65ffeb98995b3a9a1263274c978e0fcde8ed2441da9 WatchSource:0}: Error finding container a867f7b960475fce98aca65ffeb98995b3a9a1263274c978e0fcde8ed2441da9: Status 404 returned error can't find the container with id a867f7b960475fce98aca65ffeb98995b3a9a1263274c978e0fcde8ed2441da9 Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.338113 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.346365 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-klshd" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.462082 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-xbljs"] Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.521264 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.528406 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-70e2-account-create-cgqd8" Oct 02 10:06:43 crc kubenswrapper[4934]: W1002 10:06:43.567938 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda01b1d99_451c_4aec_b261_53232f25e582.slice/crio-3dbe7339f1a61bd3a78452df01c7f70bc9d2bb7d505373b48ad3bed9b996e3e0 WatchSource:0}: Error finding container 3dbe7339f1a61bd3a78452df01c7f70bc9d2bb7d505373b48ad3bed9b996e3e0: Status 404 returned error can't find the container with id 3dbe7339f1a61bd3a78452df01c7f70bc9d2bb7d505373b48ad3bed9b996e3e0 Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.626040 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2gn7\" (UniqueName: \"kubernetes.io/projected/f591dece-fc9a-430d-b0b4-5de71b477541-kube-api-access-w2gn7\") pod \"f591dece-fc9a-430d-b0b4-5de71b477541\" (UID: \"f591dece-fc9a-430d-b0b4-5de71b477541\") " Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.640241 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-kb5lq"] Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.651419 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f591dece-fc9a-430d-b0b4-5de71b477541-kube-api-access-w2gn7" (OuterVolumeSpecName: "kube-api-access-w2gn7") pod "f591dece-fc9a-430d-b0b4-5de71b477541" (UID: "f591dece-fc9a-430d-b0b4-5de71b477541"). InnerVolumeSpecName "kube-api-access-w2gn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.728064 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2gn7\" (UniqueName: \"kubernetes.io/projected/f591dece-fc9a-430d-b0b4-5de71b477541-kube-api-access-w2gn7\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.742294 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-rqm5f"] Oct 02 10:06:43 crc kubenswrapper[4934]: W1002 10:06:43.753418 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod379bc476_7f65_408d_929a_8badb1e4dfdc.slice/crio-776b050a8b5e2fb2457cecd8cb868628b610293ebb85322da88f7180123b6a94 WatchSource:0}: Error finding container 776b050a8b5e2fb2457cecd8cb868628b610293ebb85322da88f7180123b6a94: Status 404 returned error can't find the container with id 776b050a8b5e2fb2457cecd8cb868628b610293ebb85322da88f7180123b6a94 Oct 02 10:06:43 crc kubenswrapper[4934]: I1002 10:06:43.999025 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-685744b6b9-5mdh2"] Oct 02 10:06:44 crc kubenswrapper[4934]: W1002 10:06:44.004312 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2059861_459c_448a_933d_bc0d7660a9dc.slice/crio-a9b7c19121b2e90dd5c9edf16a5d9a6bae463406ba7ded086e41589b2d221c73 WatchSource:0}: Error finding container a9b7c19121b2e90dd5c9edf16a5d9a6bae463406ba7ded086e41589b2d221c73: Status 404 returned error can't find the container with id a9b7c19121b2e90dd5c9edf16a5d9a6bae463406ba7ded086e41589b2d221c73 Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.013830 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-klshd"] Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.197736 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" event={"ID":"f2059861-459c-448a-933d-bc0d7660a9dc","Type":"ContainerStarted","Data":"a9b7c19121b2e90dd5c9edf16a5d9a6bae463406ba7ded086e41589b2d221c73"} Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.199369 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-rqm5f" event={"ID":"379bc476-7f65-408d-929a-8badb1e4dfdc","Type":"ContainerStarted","Data":"776b050a8b5e2fb2457cecd8cb868628b610293ebb85322da88f7180123b6a94"} Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.200711 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xbljs" event={"ID":"6f91778c-ae60-410b-8a84-16007820b80b","Type":"ContainerStarted","Data":"0118a5fe34a7b108765ca45fbcc0401dc71fcf952a3ee604afa4074f90756118"} Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.200737 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xbljs" event={"ID":"6f91778c-ae60-410b-8a84-16007820b80b","Type":"ContainerStarted","Data":"c7238a4e2000da86b88e643aad48abf0d1baf33f1cf49b2a7185d64d890cb6df"} Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.206631 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-70e2-account-create-cgqd8" event={"ID":"f591dece-fc9a-430d-b0b4-5de71b477541","Type":"ContainerDied","Data":"b3eafddce2353e95475b8ad7ef49b05c5e4af10b9952dd93b698aecd957511bb"} Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.206670 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3eafddce2353e95475b8ad7ef49b05c5e4af10b9952dd93b698aecd957511bb" Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.206640 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-70e2-account-create-cgqd8" Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.222474 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-klshd" event={"ID":"157aa366-544e-42ba-bd81-1649575c976f","Type":"ContainerStarted","Data":"4501e43d704d24923d3a1889882ed517cc126f212d5c40ef28b2fbf054abc713"} Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.231843 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-kb5lq" event={"ID":"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb","Type":"ContainerStarted","Data":"807d9943655205ef342496d4ab704cd5e64ab362985bac33828e63b29adb19f3"} Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.233336 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-xbljs" podStartSLOduration=2.233318004 podStartE2EDuration="2.233318004s" podCreationTimestamp="2025-10-02 10:06:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:06:44.222561548 +0000 UTC m=+1075.975203090" watchObservedRunningTime="2025-10-02 10:06:44.233318004 +0000 UTC m=+1075.985959526" Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.241238 4934 generic.go:334] "Generic (PLEG): container finished" podID="be49fd0e-9182-46e3-aa25-4b2681694826" containerID="64ce8841e21fd8435a2b824f34e4ffd0e3abd8e1d40f65e6c17f063319aa9146" exitCode=0 Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.241429 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f6d4499cf-btbf8" event={"ID":"be49fd0e-9182-46e3-aa25-4b2681694826","Type":"ContainerDied","Data":"64ce8841e21fd8435a2b824f34e4ffd0e3abd8e1d40f65e6c17f063319aa9146"} Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.241456 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f6d4499cf-btbf8" event={"ID":"be49fd0e-9182-46e3-aa25-4b2681694826","Type":"ContainerStarted","Data":"a867f7b960475fce98aca65ffeb98995b3a9a1263274c978e0fcde8ed2441da9"} Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.246540 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a01b1d99-451c-4aec-b261-53232f25e582","Type":"ContainerStarted","Data":"3dbe7339f1a61bd3a78452df01c7f70bc9d2bb7d505373b48ad3bed9b996e3e0"} Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.622701 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f6d4499cf-btbf8" Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.747406 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-dns-swift-storage-0\") pod \"be49fd0e-9182-46e3-aa25-4b2681694826\" (UID: \"be49fd0e-9182-46e3-aa25-4b2681694826\") " Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.747473 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-ovsdbserver-nb\") pod \"be49fd0e-9182-46e3-aa25-4b2681694826\" (UID: \"be49fd0e-9182-46e3-aa25-4b2681694826\") " Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.747514 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-config\") pod \"be49fd0e-9182-46e3-aa25-4b2681694826\" (UID: \"be49fd0e-9182-46e3-aa25-4b2681694826\") " Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.747613 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6tjx\" (UniqueName: \"kubernetes.io/projected/be49fd0e-9182-46e3-aa25-4b2681694826-kube-api-access-v6tjx\") pod \"be49fd0e-9182-46e3-aa25-4b2681694826\" (UID: \"be49fd0e-9182-46e3-aa25-4b2681694826\") " Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.747663 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-ovsdbserver-sb\") pod \"be49fd0e-9182-46e3-aa25-4b2681694826\" (UID: \"be49fd0e-9182-46e3-aa25-4b2681694826\") " Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.747699 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-dns-svc\") pod \"be49fd0e-9182-46e3-aa25-4b2681694826\" (UID: \"be49fd0e-9182-46e3-aa25-4b2681694826\") " Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.756272 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be49fd0e-9182-46e3-aa25-4b2681694826-kube-api-access-v6tjx" (OuterVolumeSpecName: "kube-api-access-v6tjx") pod "be49fd0e-9182-46e3-aa25-4b2681694826" (UID: "be49fd0e-9182-46e3-aa25-4b2681694826"). InnerVolumeSpecName "kube-api-access-v6tjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.785404 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "be49fd0e-9182-46e3-aa25-4b2681694826" (UID: "be49fd0e-9182-46e3-aa25-4b2681694826"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.787315 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "be49fd0e-9182-46e3-aa25-4b2681694826" (UID: "be49fd0e-9182-46e3-aa25-4b2681694826"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.807465 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "be49fd0e-9182-46e3-aa25-4b2681694826" (UID: "be49fd0e-9182-46e3-aa25-4b2681694826"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.807532 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "be49fd0e-9182-46e3-aa25-4b2681694826" (UID: "be49fd0e-9182-46e3-aa25-4b2681694826"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.820751 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-config" (OuterVolumeSpecName: "config") pod "be49fd0e-9182-46e3-aa25-4b2681694826" (UID: "be49fd0e-9182-46e3-aa25-4b2681694826"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.850103 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.850149 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6tjx\" (UniqueName: \"kubernetes.io/projected/be49fd0e-9182-46e3-aa25-4b2681694826-kube-api-access-v6tjx\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.850164 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.850203 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.850212 4934 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:44 crc kubenswrapper[4934]: I1002 10:06:44.850223 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/be49fd0e-9182-46e3-aa25-4b2681694826-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:45 crc kubenswrapper[4934]: I1002 10:06:45.013497 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:06:45 crc kubenswrapper[4934]: I1002 10:06:45.256713 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f6d4499cf-btbf8" event={"ID":"be49fd0e-9182-46e3-aa25-4b2681694826","Type":"ContainerDied","Data":"a867f7b960475fce98aca65ffeb98995b3a9a1263274c978e0fcde8ed2441da9"} Oct 02 10:06:45 crc kubenswrapper[4934]: I1002 10:06:45.256762 4934 scope.go:117] "RemoveContainer" containerID="64ce8841e21fd8435a2b824f34e4ffd0e3abd8e1d40f65e6c17f063319aa9146" Oct 02 10:06:45 crc kubenswrapper[4934]: I1002 10:06:45.256871 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f6d4499cf-btbf8" Oct 02 10:06:45 crc kubenswrapper[4934]: I1002 10:06:45.259647 4934 generic.go:334] "Generic (PLEG): container finished" podID="f2059861-459c-448a-933d-bc0d7660a9dc" containerID="58914f2a2ca9fc802b30dedd381ae8e1b8d1bc5e534d9e74a0b61b4ece557ab9" exitCode=0 Oct 02 10:06:45 crc kubenswrapper[4934]: I1002 10:06:45.261700 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" event={"ID":"f2059861-459c-448a-933d-bc0d7660a9dc","Type":"ContainerDied","Data":"58914f2a2ca9fc802b30dedd381ae8e1b8d1bc5e534d9e74a0b61b4ece557ab9"} Oct 02 10:06:45 crc kubenswrapper[4934]: I1002 10:06:45.439304 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f6d4499cf-btbf8"] Oct 02 10:06:45 crc kubenswrapper[4934]: I1002 10:06:45.448411 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f6d4499cf-btbf8"] Oct 02 10:06:45 crc kubenswrapper[4934]: I1002 10:06:45.585611 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-9z9t6"] Oct 02 10:06:45 crc kubenswrapper[4934]: E1002 10:06:45.586074 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f591dece-fc9a-430d-b0b4-5de71b477541" containerName="mariadb-account-create" Oct 02 10:06:45 crc kubenswrapper[4934]: I1002 10:06:45.586091 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f591dece-fc9a-430d-b0b4-5de71b477541" containerName="mariadb-account-create" Oct 02 10:06:45 crc kubenswrapper[4934]: E1002 10:06:45.586116 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be49fd0e-9182-46e3-aa25-4b2681694826" containerName="init" Oct 02 10:06:45 crc kubenswrapper[4934]: I1002 10:06:45.586122 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="be49fd0e-9182-46e3-aa25-4b2681694826" containerName="init" Oct 02 10:06:45 crc kubenswrapper[4934]: I1002 10:06:45.586271 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="be49fd0e-9182-46e3-aa25-4b2681694826" containerName="init" Oct 02 10:06:45 crc kubenswrapper[4934]: I1002 10:06:45.586302 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f591dece-fc9a-430d-b0b4-5de71b477541" containerName="mariadb-account-create" Oct 02 10:06:45 crc kubenswrapper[4934]: I1002 10:06:45.586930 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-9z9t6" Oct 02 10:06:45 crc kubenswrapper[4934]: I1002 10:06:45.593890 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 02 10:06:45 crc kubenswrapper[4934]: I1002 10:06:45.593895 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-v9qtp" Oct 02 10:06:45 crc kubenswrapper[4934]: I1002 10:06:45.594213 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 02 10:06:45 crc kubenswrapper[4934]: I1002 10:06:45.597788 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-9z9t6"] Oct 02 10:06:45 crc kubenswrapper[4934]: I1002 10:06:45.696421 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3c03e58b-5a31-4295-ba05-d219daff78b1-config\") pod \"neutron-db-sync-9z9t6\" (UID: \"3c03e58b-5a31-4295-ba05-d219daff78b1\") " pod="openstack/neutron-db-sync-9z9t6" Oct 02 10:06:45 crc kubenswrapper[4934]: I1002 10:06:45.696470 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7ht5\" (UniqueName: \"kubernetes.io/projected/3c03e58b-5a31-4295-ba05-d219daff78b1-kube-api-access-f7ht5\") pod \"neutron-db-sync-9z9t6\" (UID: \"3c03e58b-5a31-4295-ba05-d219daff78b1\") " pod="openstack/neutron-db-sync-9z9t6" Oct 02 10:06:45 crc kubenswrapper[4934]: I1002 10:06:45.696510 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c03e58b-5a31-4295-ba05-d219daff78b1-combined-ca-bundle\") pod \"neutron-db-sync-9z9t6\" (UID: \"3c03e58b-5a31-4295-ba05-d219daff78b1\") " pod="openstack/neutron-db-sync-9z9t6" Oct 02 10:06:45 crc kubenswrapper[4934]: I1002 10:06:45.798097 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3c03e58b-5a31-4295-ba05-d219daff78b1-config\") pod \"neutron-db-sync-9z9t6\" (UID: \"3c03e58b-5a31-4295-ba05-d219daff78b1\") " pod="openstack/neutron-db-sync-9z9t6" Oct 02 10:06:45 crc kubenswrapper[4934]: I1002 10:06:45.798136 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7ht5\" (UniqueName: \"kubernetes.io/projected/3c03e58b-5a31-4295-ba05-d219daff78b1-kube-api-access-f7ht5\") pod \"neutron-db-sync-9z9t6\" (UID: \"3c03e58b-5a31-4295-ba05-d219daff78b1\") " pod="openstack/neutron-db-sync-9z9t6" Oct 02 10:06:45 crc kubenswrapper[4934]: I1002 10:06:45.798174 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c03e58b-5a31-4295-ba05-d219daff78b1-combined-ca-bundle\") pod \"neutron-db-sync-9z9t6\" (UID: \"3c03e58b-5a31-4295-ba05-d219daff78b1\") " pod="openstack/neutron-db-sync-9z9t6" Oct 02 10:06:45 crc kubenswrapper[4934]: I1002 10:06:45.814136 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3c03e58b-5a31-4295-ba05-d219daff78b1-config\") pod \"neutron-db-sync-9z9t6\" (UID: \"3c03e58b-5a31-4295-ba05-d219daff78b1\") " pod="openstack/neutron-db-sync-9z9t6" Oct 02 10:06:45 crc kubenswrapper[4934]: I1002 10:06:45.825367 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c03e58b-5a31-4295-ba05-d219daff78b1-combined-ca-bundle\") pod \"neutron-db-sync-9z9t6\" (UID: \"3c03e58b-5a31-4295-ba05-d219daff78b1\") " pod="openstack/neutron-db-sync-9z9t6" Oct 02 10:06:45 crc kubenswrapper[4934]: I1002 10:06:45.825796 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7ht5\" (UniqueName: \"kubernetes.io/projected/3c03e58b-5a31-4295-ba05-d219daff78b1-kube-api-access-f7ht5\") pod \"neutron-db-sync-9z9t6\" (UID: \"3c03e58b-5a31-4295-ba05-d219daff78b1\") " pod="openstack/neutron-db-sync-9z9t6" Oct 02 10:06:45 crc kubenswrapper[4934]: I1002 10:06:45.913323 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-9z9t6" Oct 02 10:06:46 crc kubenswrapper[4934]: I1002 10:06:46.281144 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" event={"ID":"f2059861-459c-448a-933d-bc0d7660a9dc","Type":"ContainerStarted","Data":"a2aa6bf2a4073de25d3a5fb820f9b97b8128dc4b37142dd94163513f5caff7bc"} Oct 02 10:06:46 crc kubenswrapper[4934]: I1002 10:06:46.281499 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" Oct 02 10:06:46 crc kubenswrapper[4934]: I1002 10:06:46.304257 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" podStartSLOduration=4.304241482 podStartE2EDuration="4.304241482s" podCreationTimestamp="2025-10-02 10:06:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:06:46.296238343 +0000 UTC m=+1078.048879865" watchObservedRunningTime="2025-10-02 10:06:46.304241482 +0000 UTC m=+1078.056883004" Oct 02 10:06:46 crc kubenswrapper[4934]: I1002 10:06:46.507906 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-9z9t6"] Oct 02 10:06:46 crc kubenswrapper[4934]: I1002 10:06:46.923597 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be49fd0e-9182-46e3-aa25-4b2681694826" path="/var/lib/kubelet/pods/be49fd0e-9182-46e3-aa25-4b2681694826/volumes" Oct 02 10:06:48 crc kubenswrapper[4934]: I1002 10:06:48.301523 4934 generic.go:334] "Generic (PLEG): container finished" podID="6f91778c-ae60-410b-8a84-16007820b80b" containerID="0118a5fe34a7b108765ca45fbcc0401dc71fcf952a3ee604afa4074f90756118" exitCode=0 Oct 02 10:06:48 crc kubenswrapper[4934]: I1002 10:06:48.301731 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xbljs" event={"ID":"6f91778c-ae60-410b-8a84-16007820b80b","Type":"ContainerDied","Data":"0118a5fe34a7b108765ca45fbcc0401dc71fcf952a3ee604afa4074f90756118"} Oct 02 10:06:50 crc kubenswrapper[4934]: I1002 10:06:50.947029 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xbljs" Oct 02 10:06:51 crc kubenswrapper[4934]: I1002 10:06:51.075971 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rp9rg\" (UniqueName: \"kubernetes.io/projected/6f91778c-ae60-410b-8a84-16007820b80b-kube-api-access-rp9rg\") pod \"6f91778c-ae60-410b-8a84-16007820b80b\" (UID: \"6f91778c-ae60-410b-8a84-16007820b80b\") " Oct 02 10:06:51 crc kubenswrapper[4934]: I1002 10:06:51.076121 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-credential-keys\") pod \"6f91778c-ae60-410b-8a84-16007820b80b\" (UID: \"6f91778c-ae60-410b-8a84-16007820b80b\") " Oct 02 10:06:51 crc kubenswrapper[4934]: I1002 10:06:51.076232 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-combined-ca-bundle\") pod \"6f91778c-ae60-410b-8a84-16007820b80b\" (UID: \"6f91778c-ae60-410b-8a84-16007820b80b\") " Oct 02 10:06:51 crc kubenswrapper[4934]: I1002 10:06:51.076262 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-scripts\") pod \"6f91778c-ae60-410b-8a84-16007820b80b\" (UID: \"6f91778c-ae60-410b-8a84-16007820b80b\") " Oct 02 10:06:51 crc kubenswrapper[4934]: I1002 10:06:51.076286 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-fernet-keys\") pod \"6f91778c-ae60-410b-8a84-16007820b80b\" (UID: \"6f91778c-ae60-410b-8a84-16007820b80b\") " Oct 02 10:06:51 crc kubenswrapper[4934]: I1002 10:06:51.076320 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-config-data\") pod \"6f91778c-ae60-410b-8a84-16007820b80b\" (UID: \"6f91778c-ae60-410b-8a84-16007820b80b\") " Oct 02 10:06:51 crc kubenswrapper[4934]: I1002 10:06:51.086096 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f91778c-ae60-410b-8a84-16007820b80b-kube-api-access-rp9rg" (OuterVolumeSpecName: "kube-api-access-rp9rg") pod "6f91778c-ae60-410b-8a84-16007820b80b" (UID: "6f91778c-ae60-410b-8a84-16007820b80b"). InnerVolumeSpecName "kube-api-access-rp9rg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:06:51 crc kubenswrapper[4934]: I1002 10:06:51.087183 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "6f91778c-ae60-410b-8a84-16007820b80b" (UID: "6f91778c-ae60-410b-8a84-16007820b80b"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:51 crc kubenswrapper[4934]: I1002 10:06:51.088195 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6f91778c-ae60-410b-8a84-16007820b80b" (UID: "6f91778c-ae60-410b-8a84-16007820b80b"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:51 crc kubenswrapper[4934]: I1002 10:06:51.095879 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-scripts" (OuterVolumeSpecName: "scripts") pod "6f91778c-ae60-410b-8a84-16007820b80b" (UID: "6f91778c-ae60-410b-8a84-16007820b80b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:51 crc kubenswrapper[4934]: I1002 10:06:51.105112 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-config-data" (OuterVolumeSpecName: "config-data") pod "6f91778c-ae60-410b-8a84-16007820b80b" (UID: "6f91778c-ae60-410b-8a84-16007820b80b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:51 crc kubenswrapper[4934]: I1002 10:06:51.108061 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6f91778c-ae60-410b-8a84-16007820b80b" (UID: "6f91778c-ae60-410b-8a84-16007820b80b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:06:51 crc kubenswrapper[4934]: I1002 10:06:51.179104 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:51 crc kubenswrapper[4934]: I1002 10:06:51.179153 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:51 crc kubenswrapper[4934]: I1002 10:06:51.179169 4934 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:51 crc kubenswrapper[4934]: I1002 10:06:51.179183 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:51 crc kubenswrapper[4934]: I1002 10:06:51.179198 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rp9rg\" (UniqueName: \"kubernetes.io/projected/6f91778c-ae60-410b-8a84-16007820b80b-kube-api-access-rp9rg\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:51 crc kubenswrapper[4934]: I1002 10:06:51.179214 4934 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6f91778c-ae60-410b-8a84-16007820b80b-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 02 10:06:51 crc kubenswrapper[4934]: I1002 10:06:51.343981 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-9z9t6" event={"ID":"3c03e58b-5a31-4295-ba05-d219daff78b1","Type":"ContainerStarted","Data":"76ab97f80fb3e8659dbd1a3a9d52730dc747105f159de483bc059e6fc1a45f2e"} Oct 02 10:06:51 crc kubenswrapper[4934]: I1002 10:06:51.345462 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xbljs" event={"ID":"6f91778c-ae60-410b-8a84-16007820b80b","Type":"ContainerDied","Data":"c7238a4e2000da86b88e643aad48abf0d1baf33f1cf49b2a7185d64d890cb6df"} Oct 02 10:06:51 crc kubenswrapper[4934]: I1002 10:06:51.345489 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7238a4e2000da86b88e643aad48abf0d1baf33f1cf49b2a7185d64d890cb6df" Oct 02 10:06:51 crc kubenswrapper[4934]: I1002 10:06:51.345590 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xbljs" Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.150128 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-xbljs"] Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.157175 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-xbljs"] Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.234041 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-vzw77"] Oct 02 10:06:52 crc kubenswrapper[4934]: E1002 10:06:52.234474 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f91778c-ae60-410b-8a84-16007820b80b" containerName="keystone-bootstrap" Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.234494 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f91778c-ae60-410b-8a84-16007820b80b" containerName="keystone-bootstrap" Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.234728 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f91778c-ae60-410b-8a84-16007820b80b" containerName="keystone-bootstrap" Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.235299 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-vzw77" Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.239019 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.239019 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-bkm4l" Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.239159 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.239168 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.253777 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-vzw77"] Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.304743 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-credential-keys\") pod \"keystone-bootstrap-vzw77\" (UID: \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\") " pod="openstack/keystone-bootstrap-vzw77" Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.304801 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-combined-ca-bundle\") pod \"keystone-bootstrap-vzw77\" (UID: \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\") " pod="openstack/keystone-bootstrap-vzw77" Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.304858 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfcms\" (UniqueName: \"kubernetes.io/projected/1b5b09b9-1368-4f7d-9ed2-0ba034890110-kube-api-access-wfcms\") pod \"keystone-bootstrap-vzw77\" (UID: \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\") " pod="openstack/keystone-bootstrap-vzw77" Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.304910 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-scripts\") pod \"keystone-bootstrap-vzw77\" (UID: \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\") " pod="openstack/keystone-bootstrap-vzw77" Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.304944 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-fernet-keys\") pod \"keystone-bootstrap-vzw77\" (UID: \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\") " pod="openstack/keystone-bootstrap-vzw77" Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.304977 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-config-data\") pod \"keystone-bootstrap-vzw77\" (UID: \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\") " pod="openstack/keystone-bootstrap-vzw77" Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.406377 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-scripts\") pod \"keystone-bootstrap-vzw77\" (UID: \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\") " pod="openstack/keystone-bootstrap-vzw77" Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.406440 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-fernet-keys\") pod \"keystone-bootstrap-vzw77\" (UID: \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\") " pod="openstack/keystone-bootstrap-vzw77" Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.406476 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-config-data\") pod \"keystone-bootstrap-vzw77\" (UID: \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\") " pod="openstack/keystone-bootstrap-vzw77" Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.406505 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-credential-keys\") pod \"keystone-bootstrap-vzw77\" (UID: \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\") " pod="openstack/keystone-bootstrap-vzw77" Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.406525 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-combined-ca-bundle\") pod \"keystone-bootstrap-vzw77\" (UID: \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\") " pod="openstack/keystone-bootstrap-vzw77" Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.406564 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfcms\" (UniqueName: \"kubernetes.io/projected/1b5b09b9-1368-4f7d-9ed2-0ba034890110-kube-api-access-wfcms\") pod \"keystone-bootstrap-vzw77\" (UID: \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\") " pod="openstack/keystone-bootstrap-vzw77" Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.410946 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-scripts\") pod \"keystone-bootstrap-vzw77\" (UID: \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\") " pod="openstack/keystone-bootstrap-vzw77" Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.410992 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-combined-ca-bundle\") pod \"keystone-bootstrap-vzw77\" (UID: \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\") " pod="openstack/keystone-bootstrap-vzw77" Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.411540 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-fernet-keys\") pod \"keystone-bootstrap-vzw77\" (UID: \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\") " pod="openstack/keystone-bootstrap-vzw77" Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.417246 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-credential-keys\") pod \"keystone-bootstrap-vzw77\" (UID: \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\") " pod="openstack/keystone-bootstrap-vzw77" Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.418098 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-config-data\") pod \"keystone-bootstrap-vzw77\" (UID: \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\") " pod="openstack/keystone-bootstrap-vzw77" Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.424394 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfcms\" (UniqueName: \"kubernetes.io/projected/1b5b09b9-1368-4f7d-9ed2-0ba034890110-kube-api-access-wfcms\") pod \"keystone-bootstrap-vzw77\" (UID: \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\") " pod="openstack/keystone-bootstrap-vzw77" Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.556693 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-vzw77" Oct 02 10:06:52 crc kubenswrapper[4934]: I1002 10:06:52.925237 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f91778c-ae60-410b-8a84-16007820b80b" path="/var/lib/kubelet/pods/6f91778c-ae60-410b-8a84-16007820b80b/volumes" Oct 02 10:06:53 crc kubenswrapper[4934]: I1002 10:06:53.339748 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" Oct 02 10:06:53 crc kubenswrapper[4934]: I1002 10:06:53.397984 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68d47f7449-5tf7r"] Oct 02 10:06:53 crc kubenswrapper[4934]: I1002 10:06:53.398853 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" podUID="95d46109-8486-474d-8b03-78088500da28" containerName="dnsmasq-dns" containerID="cri-o://e20b9a73c9c60067b166a4179b8e24057a49922f4c4019066a179535b25fa21c" gracePeriod=10 Oct 02 10:06:54 crc kubenswrapper[4934]: I1002 10:06:54.375939 4934 generic.go:334] "Generic (PLEG): container finished" podID="95d46109-8486-474d-8b03-78088500da28" containerID="e20b9a73c9c60067b166a4179b8e24057a49922f4c4019066a179535b25fa21c" exitCode=0 Oct 02 10:06:54 crc kubenswrapper[4934]: I1002 10:06:54.375989 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" event={"ID":"95d46109-8486-474d-8b03-78088500da28","Type":"ContainerDied","Data":"e20b9a73c9c60067b166a4179b8e24057a49922f4c4019066a179535b25fa21c"} Oct 02 10:06:55 crc kubenswrapper[4934]: I1002 10:06:55.659471 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" podUID="95d46109-8486-474d-8b03-78088500da28" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.131:5353: connect: connection refused" Oct 02 10:07:00 crc kubenswrapper[4934]: I1002 10:07:00.660073 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" podUID="95d46109-8486-474d-8b03-78088500da28" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.131:5353: connect: connection refused" Oct 02 10:07:04 crc kubenswrapper[4934]: E1002 10:07:04.157179 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api@sha256:d88750a3464a2b6dd3bfcfc7222a9579b2aebdf23014ce835ad0b0d8492d0ad9" Oct 02 10:07:04 crc kubenswrapper[4934]: E1002 10:07:04.157969 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api@sha256:d88750a3464a2b6dd3bfcfc7222a9579b2aebdf23014ce835ad0b0d8492d0ad9,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4jp9h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-klshd_openstack(157aa366-544e-42ba-bd81-1649575c976f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 10:07:04 crc kubenswrapper[4934]: E1002 10:07:04.159264 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-klshd" podUID="157aa366-544e-42ba-bd81-1649575c976f" Oct 02 10:07:04 crc kubenswrapper[4934]: E1002 10:07:04.465561 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api@sha256:d88750a3464a2b6dd3bfcfc7222a9579b2aebdf23014ce835ad0b0d8492d0ad9\\\"\"" pod="openstack/placement-db-sync-klshd" podUID="157aa366-544e-42ba-bd81-1649575c976f" Oct 02 10:07:05 crc kubenswrapper[4934]: I1002 10:07:05.659092 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" podUID="95d46109-8486-474d-8b03-78088500da28" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.131:5353: connect: connection refused" Oct 02 10:07:05 crc kubenswrapper[4934]: I1002 10:07:05.659189 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" Oct 02 10:07:08 crc kubenswrapper[4934]: E1002 10:07:08.259324 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:dec0378fb73f66dd39f6d7f7bdc6020d9bdf4c57c201eac3c63d7e97667627fe" Oct 02 10:07:08 crc kubenswrapper[4934]: E1002 10:07:08.260286 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:dec0378fb73f66dd39f6d7f7bdc6020d9bdf4c57c201eac3c63d7e97667627fe,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qlskg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-kb5lq_openstack(087ad031-cdf0-45e4-b1cf-ae777e8cc5fb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 10:07:08 crc kubenswrapper[4934]: E1002 10:07:08.261523 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-kb5lq" podUID="087ad031-cdf0-45e4-b1cf-ae777e8cc5fb" Oct 02 10:07:08 crc kubenswrapper[4934]: I1002 10:07:08.440381 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:07:08 crc kubenswrapper[4934]: I1002 10:07:08.440712 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:07:08 crc kubenswrapper[4934]: E1002 10:07:08.502216 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:dec0378fb73f66dd39f6d7f7bdc6020d9bdf4c57c201eac3c63d7e97667627fe\\\"\"" pod="openstack/cinder-db-sync-kb5lq" podUID="087ad031-cdf0-45e4-b1cf-ae777e8cc5fb" Oct 02 10:07:08 crc kubenswrapper[4934]: E1002 10:07:08.886221 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api@sha256:85c8d86100270d60e99d2a7ab5e875fa7634a6a8a6c351630fe3b964e1b11f0e" Oct 02 10:07:08 crc kubenswrapper[4934]: E1002 10:07:08.886462 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:85c8d86100270d60e99d2a7ab5e875fa7634a6a8a6c351630fe3b964e1b11f0e,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zqq9t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-jwpkw_openstack(6427286d-a1ca-43ca-8ecb-bbdda45e360e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 10:07:08 crc kubenswrapper[4934]: E1002 10:07:08.887696 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-jwpkw" podUID="6427286d-a1ca-43ca-8ecb-bbdda45e360e" Oct 02 10:07:08 crc kubenswrapper[4934]: E1002 10:07:08.928369 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:4da48d8f16bfe4f35784b91b85bbd936c35ed26274a3991c54a19e6520c41c3c" Oct 02 10:07:08 crc kubenswrapper[4934]: E1002 10:07:08.928923 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:4da48d8f16bfe4f35784b91b85bbd936c35ed26274a3991c54a19e6520c41c3c,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bmzvx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-rqm5f_openstack(379bc476-7f65-408d-929a-8badb1e4dfdc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 10:07:08 crc kubenswrapper[4934]: E1002 10:07:08.930116 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-rqm5f" podUID="379bc476-7f65-408d-929a-8badb1e4dfdc" Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.051092 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.121723 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-dns-svc\") pod \"95d46109-8486-474d-8b03-78088500da28\" (UID: \"95d46109-8486-474d-8b03-78088500da28\") " Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.122219 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-ovsdbserver-sb\") pod \"95d46109-8486-474d-8b03-78088500da28\" (UID: \"95d46109-8486-474d-8b03-78088500da28\") " Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.122257 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trnj8\" (UniqueName: \"kubernetes.io/projected/95d46109-8486-474d-8b03-78088500da28-kube-api-access-trnj8\") pod \"95d46109-8486-474d-8b03-78088500da28\" (UID: \"95d46109-8486-474d-8b03-78088500da28\") " Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.122303 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-ovsdbserver-nb\") pod \"95d46109-8486-474d-8b03-78088500da28\" (UID: \"95d46109-8486-474d-8b03-78088500da28\") " Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.122336 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-config\") pod \"95d46109-8486-474d-8b03-78088500da28\" (UID: \"95d46109-8486-474d-8b03-78088500da28\") " Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.122407 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-dns-swift-storage-0\") pod \"95d46109-8486-474d-8b03-78088500da28\" (UID: \"95d46109-8486-474d-8b03-78088500da28\") " Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.138276 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95d46109-8486-474d-8b03-78088500da28-kube-api-access-trnj8" (OuterVolumeSpecName: "kube-api-access-trnj8") pod "95d46109-8486-474d-8b03-78088500da28" (UID: "95d46109-8486-474d-8b03-78088500da28"). InnerVolumeSpecName "kube-api-access-trnj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.182227 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "95d46109-8486-474d-8b03-78088500da28" (UID: "95d46109-8486-474d-8b03-78088500da28"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.182621 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-config" (OuterVolumeSpecName: "config") pod "95d46109-8486-474d-8b03-78088500da28" (UID: "95d46109-8486-474d-8b03-78088500da28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.187812 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "95d46109-8486-474d-8b03-78088500da28" (UID: "95d46109-8486-474d-8b03-78088500da28"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.190173 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "95d46109-8486-474d-8b03-78088500da28" (UID: "95d46109-8486-474d-8b03-78088500da28"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.205254 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "95d46109-8486-474d-8b03-78088500da28" (UID: "95d46109-8486-474d-8b03-78088500da28"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.224269 4934 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.224317 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.224331 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.224343 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trnj8\" (UniqueName: \"kubernetes.io/projected/95d46109-8486-474d-8b03-78088500da28-kube-api-access-trnj8\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.224358 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.224368 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95d46109-8486-474d-8b03-78088500da28-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.383729 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-vzw77"] Oct 02 10:07:09 crc kubenswrapper[4934]: W1002 10:07:09.386566 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b5b09b9_1368_4f7d_9ed2_0ba034890110.slice/crio-4a0674dcbae6e2314d1c99b85d8d1c0a350458d50ef26c91a0a01d80fe63a007 WatchSource:0}: Error finding container 4a0674dcbae6e2314d1c99b85d8d1c0a350458d50ef26c91a0a01d80fe63a007: Status 404 returned error can't find the container with id 4a0674dcbae6e2314d1c99b85d8d1c0a350458d50ef26c91a0a01d80fe63a007 Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.508624 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a01b1d99-451c-4aec-b261-53232f25e582","Type":"ContainerStarted","Data":"002f687aab79998010a025265fe48f45b2877e85522eee84029b12d3e0ae5b91"} Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.510058 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-vzw77" event={"ID":"1b5b09b9-1368-4f7d-9ed2-0ba034890110","Type":"ContainerStarted","Data":"4a0674dcbae6e2314d1c99b85d8d1c0a350458d50ef26c91a0a01d80fe63a007"} Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.515089 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-9z9t6" event={"ID":"3c03e58b-5a31-4295-ba05-d219daff78b1","Type":"ContainerStarted","Data":"fd0b5910cc8779a765c024bbeafee2d307597efe1d65c9d3dbff30dc96cca647"} Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.525838 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.525849 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d47f7449-5tf7r" event={"ID":"95d46109-8486-474d-8b03-78088500da28","Type":"ContainerDied","Data":"5ba0bc49fe537e10fe256fd5be21a999522b603a1bb47fbfe679c6f97ee5fafa"} Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.525929 4934 scope.go:117] "RemoveContainer" containerID="e20b9a73c9c60067b166a4179b8e24057a49922f4c4019066a179535b25fa21c" Oct 02 10:07:09 crc kubenswrapper[4934]: E1002 10:07:09.538233 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:4da48d8f16bfe4f35784b91b85bbd936c35ed26274a3991c54a19e6520c41c3c\\\"\"" pod="openstack/barbican-db-sync-rqm5f" podUID="379bc476-7f65-408d-929a-8badb1e4dfdc" Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.546113 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-9z9t6" podStartSLOduration=24.546085574 podStartE2EDuration="24.546085574s" podCreationTimestamp="2025-10-02 10:06:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:07:09.532455586 +0000 UTC m=+1101.285097108" watchObservedRunningTime="2025-10-02 10:07:09.546085574 +0000 UTC m=+1101.298727106" Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.548932 4934 scope.go:117] "RemoveContainer" containerID="97c0006b77d839328731430ad86d47bb8ca5d18e833d3d3a511d11adab1fd1df" Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.582241 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68d47f7449-5tf7r"] Oct 02 10:07:09 crc kubenswrapper[4934]: I1002 10:07:09.593140 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-68d47f7449-5tf7r"] Oct 02 10:07:10 crc kubenswrapper[4934]: I1002 10:07:10.555204 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-vzw77" event={"ID":"1b5b09b9-1368-4f7d-9ed2-0ba034890110","Type":"ContainerStarted","Data":"2685a9eb90de41e74a709a47f4f8313728ce1ab7fcb3df49f017512f156c070d"} Oct 02 10:07:10 crc kubenswrapper[4934]: I1002 10:07:10.575023 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-vzw77" podStartSLOduration=18.574980111 podStartE2EDuration="18.574980111s" podCreationTimestamp="2025-10-02 10:06:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:07:10.569363281 +0000 UTC m=+1102.322004813" watchObservedRunningTime="2025-10-02 10:07:10.574980111 +0000 UTC m=+1102.327621633" Oct 02 10:07:10 crc kubenswrapper[4934]: I1002 10:07:10.923608 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95d46109-8486-474d-8b03-78088500da28" path="/var/lib/kubelet/pods/95d46109-8486-474d-8b03-78088500da28/volumes" Oct 02 10:07:11 crc kubenswrapper[4934]: I1002 10:07:11.567812 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a01b1d99-451c-4aec-b261-53232f25e582","Type":"ContainerStarted","Data":"36e7f69a1fab279fde6bbae278862639f78714530357b94feba58581be27ce11"} Oct 02 10:07:14 crc kubenswrapper[4934]: I1002 10:07:14.640992 4934 generic.go:334] "Generic (PLEG): container finished" podID="1b5b09b9-1368-4f7d-9ed2-0ba034890110" containerID="2685a9eb90de41e74a709a47f4f8313728ce1ab7fcb3df49f017512f156c070d" exitCode=0 Oct 02 10:07:14 crc kubenswrapper[4934]: I1002 10:07:14.641332 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-vzw77" event={"ID":"1b5b09b9-1368-4f7d-9ed2-0ba034890110","Type":"ContainerDied","Data":"2685a9eb90de41e74a709a47f4f8313728ce1ab7fcb3df49f017512f156c070d"} Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.301437 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-vzw77" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.359240 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfcms\" (UniqueName: \"kubernetes.io/projected/1b5b09b9-1368-4f7d-9ed2-0ba034890110-kube-api-access-wfcms\") pod \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\" (UID: \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\") " Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.359302 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-fernet-keys\") pod \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\" (UID: \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\") " Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.359340 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-combined-ca-bundle\") pod \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\" (UID: \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\") " Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.359434 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-credential-keys\") pod \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\" (UID: \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\") " Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.359470 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-scripts\") pod \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\" (UID: \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\") " Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.359487 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-config-data\") pod \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\" (UID: \"1b5b09b9-1368-4f7d-9ed2-0ba034890110\") " Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.369494 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "1b5b09b9-1368-4f7d-9ed2-0ba034890110" (UID: "1b5b09b9-1368-4f7d-9ed2-0ba034890110"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.369624 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b5b09b9-1368-4f7d-9ed2-0ba034890110-kube-api-access-wfcms" (OuterVolumeSpecName: "kube-api-access-wfcms") pod "1b5b09b9-1368-4f7d-9ed2-0ba034890110" (UID: "1b5b09b9-1368-4f7d-9ed2-0ba034890110"). InnerVolumeSpecName "kube-api-access-wfcms". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.375773 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-scripts" (OuterVolumeSpecName: "scripts") pod "1b5b09b9-1368-4f7d-9ed2-0ba034890110" (UID: "1b5b09b9-1368-4f7d-9ed2-0ba034890110"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.376445 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "1b5b09b9-1368-4f7d-9ed2-0ba034890110" (UID: "1b5b09b9-1368-4f7d-9ed2-0ba034890110"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.397391 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1b5b09b9-1368-4f7d-9ed2-0ba034890110" (UID: "1b5b09b9-1368-4f7d-9ed2-0ba034890110"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.400222 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-config-data" (OuterVolumeSpecName: "config-data") pod "1b5b09b9-1368-4f7d-9ed2-0ba034890110" (UID: "1b5b09b9-1368-4f7d-9ed2-0ba034890110"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.460926 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.460963 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.460998 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfcms\" (UniqueName: \"kubernetes.io/projected/1b5b09b9-1368-4f7d-9ed2-0ba034890110-kube-api-access-wfcms\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.461009 4934 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.461017 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.461025 4934 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1b5b09b9-1368-4f7d-9ed2-0ba034890110-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.665077 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a01b1d99-451c-4aec-b261-53232f25e582","Type":"ContainerStarted","Data":"6b06bbdf3c00a0bc0c714c9fa3f62f90c4ca4d7277e674e75b54eb86e9ba083a"} Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.667060 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-vzw77" event={"ID":"1b5b09b9-1368-4f7d-9ed2-0ba034890110","Type":"ContainerDied","Data":"4a0674dcbae6e2314d1c99b85d8d1c0a350458d50ef26c91a0a01d80fe63a007"} Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.667084 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a0674dcbae6e2314d1c99b85d8d1c0a350458d50ef26c91a0a01d80fe63a007" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.667238 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-vzw77" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.815002 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7b7b4b7b85-9llg8"] Oct 02 10:07:16 crc kubenswrapper[4934]: E1002 10:07:16.815648 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b5b09b9-1368-4f7d-9ed2-0ba034890110" containerName="keystone-bootstrap" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.815747 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b5b09b9-1368-4f7d-9ed2-0ba034890110" containerName="keystone-bootstrap" Oct 02 10:07:16 crc kubenswrapper[4934]: E1002 10:07:16.815831 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95d46109-8486-474d-8b03-78088500da28" containerName="init" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.815893 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="95d46109-8486-474d-8b03-78088500da28" containerName="init" Oct 02 10:07:16 crc kubenswrapper[4934]: E1002 10:07:16.815976 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95d46109-8486-474d-8b03-78088500da28" containerName="dnsmasq-dns" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.816037 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="95d46109-8486-474d-8b03-78088500da28" containerName="dnsmasq-dns" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.816302 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="95d46109-8486-474d-8b03-78088500da28" containerName="dnsmasq-dns" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.816392 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b5b09b9-1368-4f7d-9ed2-0ba034890110" containerName="keystone-bootstrap" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.817059 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.819980 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.820286 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.820759 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.821803 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.822277 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-bkm4l" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.824969 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.846143 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7b7b4b7b85-9llg8"] Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.867836 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-credential-keys\") pod \"keystone-7b7b4b7b85-9llg8\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.867891 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-public-tls-certs\") pod \"keystone-7b7b4b7b85-9llg8\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.867941 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-fernet-keys\") pod \"keystone-7b7b4b7b85-9llg8\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.867969 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb5p5\" (UniqueName: \"kubernetes.io/projected/02feb54e-6b14-42ec-a22d-524a6005e2fd-kube-api-access-bb5p5\") pod \"keystone-7b7b4b7b85-9llg8\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.867997 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-scripts\") pod \"keystone-7b7b4b7b85-9llg8\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.868016 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-combined-ca-bundle\") pod \"keystone-7b7b4b7b85-9llg8\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.868097 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-internal-tls-certs\") pod \"keystone-7b7b4b7b85-9llg8\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.868123 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-config-data\") pod \"keystone-7b7b4b7b85-9llg8\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.969256 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-internal-tls-certs\") pod \"keystone-7b7b4b7b85-9llg8\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.969301 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-config-data\") pod \"keystone-7b7b4b7b85-9llg8\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.969397 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-credential-keys\") pod \"keystone-7b7b4b7b85-9llg8\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.969419 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-public-tls-certs\") pod \"keystone-7b7b4b7b85-9llg8\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.969450 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-fernet-keys\") pod \"keystone-7b7b4b7b85-9llg8\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.969494 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb5p5\" (UniqueName: \"kubernetes.io/projected/02feb54e-6b14-42ec-a22d-524a6005e2fd-kube-api-access-bb5p5\") pod \"keystone-7b7b4b7b85-9llg8\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.969514 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-scripts\") pod \"keystone-7b7b4b7b85-9llg8\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.969529 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-combined-ca-bundle\") pod \"keystone-7b7b4b7b85-9llg8\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.973493 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-internal-tls-certs\") pod \"keystone-7b7b4b7b85-9llg8\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.973917 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-credential-keys\") pod \"keystone-7b7b4b7b85-9llg8\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.974089 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-public-tls-certs\") pod \"keystone-7b7b4b7b85-9llg8\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.974514 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-config-data\") pod \"keystone-7b7b4b7b85-9llg8\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.975189 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-combined-ca-bundle\") pod \"keystone-7b7b4b7b85-9llg8\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.977142 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-fernet-keys\") pod \"keystone-7b7b4b7b85-9llg8\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.980908 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-scripts\") pod \"keystone-7b7b4b7b85-9llg8\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:07:16 crc kubenswrapper[4934]: I1002 10:07:16.987817 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb5p5\" (UniqueName: \"kubernetes.io/projected/02feb54e-6b14-42ec-a22d-524a6005e2fd-kube-api-access-bb5p5\") pod \"keystone-7b7b4b7b85-9llg8\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:07:17 crc kubenswrapper[4934]: I1002 10:07:17.131306 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:07:17 crc kubenswrapper[4934]: I1002 10:07:17.614505 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7b7b4b7b85-9llg8"] Oct 02 10:07:17 crc kubenswrapper[4934]: I1002 10:07:17.677438 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7b7b4b7b85-9llg8" event={"ID":"02feb54e-6b14-42ec-a22d-524a6005e2fd","Type":"ContainerStarted","Data":"499eff26242dc376d8606b745f5ee91a43e209b362e362587d1c1628a3a92c05"} Oct 02 10:07:18 crc kubenswrapper[4934]: I1002 10:07:18.687757 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7b7b4b7b85-9llg8" event={"ID":"02feb54e-6b14-42ec-a22d-524a6005e2fd","Type":"ContainerStarted","Data":"e7f1b0e43bf1c9bf7da08e7461ca997f5bf6711c2a32f9d4be458d1aff9e7cdc"} Oct 02 10:07:18 crc kubenswrapper[4934]: I1002 10:07:18.688091 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:07:18 crc kubenswrapper[4934]: I1002 10:07:18.707272 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7b7b4b7b85-9llg8" podStartSLOduration=2.707229094 podStartE2EDuration="2.707229094s" podCreationTimestamp="2025-10-02 10:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:07:18.703603931 +0000 UTC m=+1110.456245453" watchObservedRunningTime="2025-10-02 10:07:18.707229094 +0000 UTC m=+1110.459870616" Oct 02 10:07:22 crc kubenswrapper[4934]: E1002 10:07:22.921060 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api@sha256:85c8d86100270d60e99d2a7ab5e875fa7634a6a8a6c351630fe3b964e1b11f0e\\\"\"" pod="openstack/glance-db-sync-jwpkw" podUID="6427286d-a1ca-43ca-8ecb-bbdda45e360e" Oct 02 10:07:31 crc kubenswrapper[4934]: I1002 10:07:31.806188 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-rqm5f" event={"ID":"379bc476-7f65-408d-929a-8badb1e4dfdc","Type":"ContainerStarted","Data":"d4c19f695157dd25d239073bba217146b7e1c30d74c166a1697941fd5eab8a53"} Oct 02 10:07:31 crc kubenswrapper[4934]: I1002 10:07:31.809171 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-klshd" event={"ID":"157aa366-544e-42ba-bd81-1649575c976f","Type":"ContainerStarted","Data":"9d35a7ad52afadce99f8b4bb3bb5a4238db85b29282f3f5f2fa42c6bd9b238ef"} Oct 02 10:07:31 crc kubenswrapper[4934]: I1002 10:07:31.830119 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-rqm5f" podStartSLOduration=2.379579294 podStartE2EDuration="49.830103516s" podCreationTimestamp="2025-10-02 10:06:42 +0000 UTC" firstStartedPulling="2025-10-02 10:06:43.756290498 +0000 UTC m=+1075.508932020" lastFinishedPulling="2025-10-02 10:07:31.20681472 +0000 UTC m=+1122.959456242" observedRunningTime="2025-10-02 10:07:31.824616569 +0000 UTC m=+1123.577258091" watchObservedRunningTime="2025-10-02 10:07:31.830103516 +0000 UTC m=+1123.582745038" Oct 02 10:07:31 crc kubenswrapper[4934]: I1002 10:07:31.861282 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-klshd" podStartSLOduration=2.6723640189999998 podStartE2EDuration="49.861259174s" podCreationTimestamp="2025-10-02 10:06:42 +0000 UTC" firstStartedPulling="2025-10-02 10:06:44.017964156 +0000 UTC m=+1075.770605678" lastFinishedPulling="2025-10-02 10:07:31.206859311 +0000 UTC m=+1122.959500833" observedRunningTime="2025-10-02 10:07:31.857121675 +0000 UTC m=+1123.609763197" watchObservedRunningTime="2025-10-02 10:07:31.861259174 +0000 UTC m=+1123.613900706" Oct 02 10:07:34 crc kubenswrapper[4934]: I1002 10:07:34.840211 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a01b1d99-451c-4aec-b261-53232f25e582","Type":"ContainerStarted","Data":"a54cc62ec161119238e993569676671317a5b21ee119e06686939303357c9cf1"} Oct 02 10:07:34 crc kubenswrapper[4934]: I1002 10:07:34.840634 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 10:07:34 crc kubenswrapper[4934]: I1002 10:07:34.840359 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a01b1d99-451c-4aec-b261-53232f25e582" containerName="ceilometer-central-agent" containerID="cri-o://002f687aab79998010a025265fe48f45b2877e85522eee84029b12d3e0ae5b91" gracePeriod=30 Oct 02 10:07:34 crc kubenswrapper[4934]: I1002 10:07:34.840816 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a01b1d99-451c-4aec-b261-53232f25e582" containerName="sg-core" containerID="cri-o://6b06bbdf3c00a0bc0c714c9fa3f62f90c4ca4d7277e674e75b54eb86e9ba083a" gracePeriod=30 Oct 02 10:07:34 crc kubenswrapper[4934]: I1002 10:07:34.840818 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a01b1d99-451c-4aec-b261-53232f25e582" containerName="proxy-httpd" containerID="cri-o://a54cc62ec161119238e993569676671317a5b21ee119e06686939303357c9cf1" gracePeriod=30 Oct 02 10:07:34 crc kubenswrapper[4934]: I1002 10:07:34.840903 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a01b1d99-451c-4aec-b261-53232f25e582" containerName="ceilometer-notification-agent" containerID="cri-o://36e7f69a1fab279fde6bbae278862639f78714530357b94feba58581be27ce11" gracePeriod=30 Oct 02 10:07:34 crc kubenswrapper[4934]: I1002 10:07:34.863066 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.1363109 podStartE2EDuration="52.863046623s" podCreationTimestamp="2025-10-02 10:06:42 +0000 UTC" firstStartedPulling="2025-10-02 10:06:43.573733065 +0000 UTC m=+1075.326374587" lastFinishedPulling="2025-10-02 10:07:34.300468788 +0000 UTC m=+1126.053110310" observedRunningTime="2025-10-02 10:07:34.857735782 +0000 UTC m=+1126.610377304" watchObservedRunningTime="2025-10-02 10:07:34.863046623 +0000 UTC m=+1126.615688155" Oct 02 10:07:35 crc kubenswrapper[4934]: I1002 10:07:35.852909 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-kb5lq" event={"ID":"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb","Type":"ContainerStarted","Data":"1cd0dc5312e4c3b419afaad3821f775ed69083d11a25939daa211e3938162c93"} Oct 02 10:07:35 crc kubenswrapper[4934]: I1002 10:07:35.866800 4934 generic.go:334] "Generic (PLEG): container finished" podID="a01b1d99-451c-4aec-b261-53232f25e582" containerID="a54cc62ec161119238e993569676671317a5b21ee119e06686939303357c9cf1" exitCode=0 Oct 02 10:07:35 crc kubenswrapper[4934]: I1002 10:07:35.866839 4934 generic.go:334] "Generic (PLEG): container finished" podID="a01b1d99-451c-4aec-b261-53232f25e582" containerID="6b06bbdf3c00a0bc0c714c9fa3f62f90c4ca4d7277e674e75b54eb86e9ba083a" exitCode=2 Oct 02 10:07:35 crc kubenswrapper[4934]: I1002 10:07:35.866850 4934 generic.go:334] "Generic (PLEG): container finished" podID="a01b1d99-451c-4aec-b261-53232f25e582" containerID="36e7f69a1fab279fde6bbae278862639f78714530357b94feba58581be27ce11" exitCode=0 Oct 02 10:07:35 crc kubenswrapper[4934]: I1002 10:07:35.866858 4934 generic.go:334] "Generic (PLEG): container finished" podID="a01b1d99-451c-4aec-b261-53232f25e582" containerID="002f687aab79998010a025265fe48f45b2877e85522eee84029b12d3e0ae5b91" exitCode=0 Oct 02 10:07:35 crc kubenswrapper[4934]: I1002 10:07:35.866948 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a01b1d99-451c-4aec-b261-53232f25e582","Type":"ContainerDied","Data":"a54cc62ec161119238e993569676671317a5b21ee119e06686939303357c9cf1"} Oct 02 10:07:35 crc kubenswrapper[4934]: I1002 10:07:35.866976 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a01b1d99-451c-4aec-b261-53232f25e582","Type":"ContainerDied","Data":"6b06bbdf3c00a0bc0c714c9fa3f62f90c4ca4d7277e674e75b54eb86e9ba083a"} Oct 02 10:07:35 crc kubenswrapper[4934]: I1002 10:07:35.866986 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a01b1d99-451c-4aec-b261-53232f25e582","Type":"ContainerDied","Data":"36e7f69a1fab279fde6bbae278862639f78714530357b94feba58581be27ce11"} Oct 02 10:07:35 crc kubenswrapper[4934]: I1002 10:07:35.866997 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a01b1d99-451c-4aec-b261-53232f25e582","Type":"ContainerDied","Data":"002f687aab79998010a025265fe48f45b2877e85522eee84029b12d3e0ae5b91"} Oct 02 10:07:35 crc kubenswrapper[4934]: I1002 10:07:35.875158 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jwpkw" event={"ID":"6427286d-a1ca-43ca-8ecb-bbdda45e360e","Type":"ContainerStarted","Data":"d4c05b6d8b3d3381710c9dd68250fc8af41ac47622cf6da8cc3795965ae20a68"} Oct 02 10:07:35 crc kubenswrapper[4934]: I1002 10:07:35.886666 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-kb5lq" podStartSLOduration=3.242355516 podStartE2EDuration="53.88664987s" podCreationTimestamp="2025-10-02 10:06:42 +0000 UTC" firstStartedPulling="2025-10-02 10:06:43.657151112 +0000 UTC m=+1075.409792634" lastFinishedPulling="2025-10-02 10:07:34.301445446 +0000 UTC m=+1126.054086988" observedRunningTime="2025-10-02 10:07:35.881634486 +0000 UTC m=+1127.634276008" watchObservedRunningTime="2025-10-02 10:07:35.88664987 +0000 UTC m=+1127.639291392" Oct 02 10:07:35 crc kubenswrapper[4934]: I1002 10:07:35.906054 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-jwpkw" podStartSLOduration=2.441415811 podStartE2EDuration="1m14.906033751s" podCreationTimestamp="2025-10-02 10:06:21 +0000 UTC" firstStartedPulling="2025-10-02 10:06:22.717164128 +0000 UTC m=+1054.469805660" lastFinishedPulling="2025-10-02 10:07:35.181782068 +0000 UTC m=+1126.934423600" observedRunningTime="2025-10-02 10:07:35.903157859 +0000 UTC m=+1127.655799381" watchObservedRunningTime="2025-10-02 10:07:35.906033751 +0000 UTC m=+1127.658675273" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.060655 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.174538 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a01b1d99-451c-4aec-b261-53232f25e582-scripts\") pod \"a01b1d99-451c-4aec-b261-53232f25e582\" (UID: \"a01b1d99-451c-4aec-b261-53232f25e582\") " Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.174797 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a01b1d99-451c-4aec-b261-53232f25e582-log-httpd\") pod \"a01b1d99-451c-4aec-b261-53232f25e582\" (UID: \"a01b1d99-451c-4aec-b261-53232f25e582\") " Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.174923 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a01b1d99-451c-4aec-b261-53232f25e582-config-data\") pod \"a01b1d99-451c-4aec-b261-53232f25e582\" (UID: \"a01b1d99-451c-4aec-b261-53232f25e582\") " Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.175022 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a01b1d99-451c-4aec-b261-53232f25e582-combined-ca-bundle\") pod \"a01b1d99-451c-4aec-b261-53232f25e582\" (UID: \"a01b1d99-451c-4aec-b261-53232f25e582\") " Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.175138 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qtm6\" (UniqueName: \"kubernetes.io/projected/a01b1d99-451c-4aec-b261-53232f25e582-kube-api-access-5qtm6\") pod \"a01b1d99-451c-4aec-b261-53232f25e582\" (UID: \"a01b1d99-451c-4aec-b261-53232f25e582\") " Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.175223 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a01b1d99-451c-4aec-b261-53232f25e582-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a01b1d99-451c-4aec-b261-53232f25e582" (UID: "a01b1d99-451c-4aec-b261-53232f25e582"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.175335 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a01b1d99-451c-4aec-b261-53232f25e582-run-httpd\") pod \"a01b1d99-451c-4aec-b261-53232f25e582\" (UID: \"a01b1d99-451c-4aec-b261-53232f25e582\") " Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.175489 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a01b1d99-451c-4aec-b261-53232f25e582-sg-core-conf-yaml\") pod \"a01b1d99-451c-4aec-b261-53232f25e582\" (UID: \"a01b1d99-451c-4aec-b261-53232f25e582\") " Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.175555 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a01b1d99-451c-4aec-b261-53232f25e582-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a01b1d99-451c-4aec-b261-53232f25e582" (UID: "a01b1d99-451c-4aec-b261-53232f25e582"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.175919 4934 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a01b1d99-451c-4aec-b261-53232f25e582-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.175987 4934 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a01b1d99-451c-4aec-b261-53232f25e582-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.180302 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a01b1d99-451c-4aec-b261-53232f25e582-kube-api-access-5qtm6" (OuterVolumeSpecName: "kube-api-access-5qtm6") pod "a01b1d99-451c-4aec-b261-53232f25e582" (UID: "a01b1d99-451c-4aec-b261-53232f25e582"). InnerVolumeSpecName "kube-api-access-5qtm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.180768 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a01b1d99-451c-4aec-b261-53232f25e582-scripts" (OuterVolumeSpecName: "scripts") pod "a01b1d99-451c-4aec-b261-53232f25e582" (UID: "a01b1d99-451c-4aec-b261-53232f25e582"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.200871 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a01b1d99-451c-4aec-b261-53232f25e582-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a01b1d99-451c-4aec-b261-53232f25e582" (UID: "a01b1d99-451c-4aec-b261-53232f25e582"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.246208 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a01b1d99-451c-4aec-b261-53232f25e582-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a01b1d99-451c-4aec-b261-53232f25e582" (UID: "a01b1d99-451c-4aec-b261-53232f25e582"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.288765 4934 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a01b1d99-451c-4aec-b261-53232f25e582-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.288806 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a01b1d99-451c-4aec-b261-53232f25e582-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.288819 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a01b1d99-451c-4aec-b261-53232f25e582-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.288832 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qtm6\" (UniqueName: \"kubernetes.io/projected/a01b1d99-451c-4aec-b261-53232f25e582-kube-api-access-5qtm6\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.329750 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a01b1d99-451c-4aec-b261-53232f25e582-config-data" (OuterVolumeSpecName: "config-data") pod "a01b1d99-451c-4aec-b261-53232f25e582" (UID: "a01b1d99-451c-4aec-b261-53232f25e582"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.393612 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a01b1d99-451c-4aec-b261-53232f25e582-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.889049 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a01b1d99-451c-4aec-b261-53232f25e582","Type":"ContainerDied","Data":"3dbe7339f1a61bd3a78452df01c7f70bc9d2bb7d505373b48ad3bed9b996e3e0"} Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.889485 4934 scope.go:117] "RemoveContainer" containerID="a54cc62ec161119238e993569676671317a5b21ee119e06686939303357c9cf1" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.889989 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.940163 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.940742 4934 scope.go:117] "RemoveContainer" containerID="6b06bbdf3c00a0bc0c714c9fa3f62f90c4ca4d7277e674e75b54eb86e9ba083a" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.954552 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.963887 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:07:36 crc kubenswrapper[4934]: E1002 10:07:36.964308 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a01b1d99-451c-4aec-b261-53232f25e582" containerName="sg-core" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.964329 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a01b1d99-451c-4aec-b261-53232f25e582" containerName="sg-core" Oct 02 10:07:36 crc kubenswrapper[4934]: E1002 10:07:36.964360 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a01b1d99-451c-4aec-b261-53232f25e582" containerName="proxy-httpd" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.964369 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a01b1d99-451c-4aec-b261-53232f25e582" containerName="proxy-httpd" Oct 02 10:07:36 crc kubenswrapper[4934]: E1002 10:07:36.964390 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a01b1d99-451c-4aec-b261-53232f25e582" containerName="ceilometer-central-agent" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.964397 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a01b1d99-451c-4aec-b261-53232f25e582" containerName="ceilometer-central-agent" Oct 02 10:07:36 crc kubenswrapper[4934]: E1002 10:07:36.964407 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a01b1d99-451c-4aec-b261-53232f25e582" containerName="ceilometer-notification-agent" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.964414 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a01b1d99-451c-4aec-b261-53232f25e582" containerName="ceilometer-notification-agent" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.964911 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="a01b1d99-451c-4aec-b261-53232f25e582" containerName="ceilometer-central-agent" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.964947 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="a01b1d99-451c-4aec-b261-53232f25e582" containerName="ceilometer-notification-agent" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.964965 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="a01b1d99-451c-4aec-b261-53232f25e582" containerName="proxy-httpd" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.964994 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="a01b1d99-451c-4aec-b261-53232f25e582" containerName="sg-core" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.967558 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.970760 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.970849 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.983695 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:07:36 crc kubenswrapper[4934]: I1002 10:07:36.999569 4934 scope.go:117] "RemoveContainer" containerID="36e7f69a1fab279fde6bbae278862639f78714530357b94feba58581be27ce11" Oct 02 10:07:37 crc kubenswrapper[4934]: I1002 10:07:37.030926 4934 scope.go:117] "RemoveContainer" containerID="002f687aab79998010a025265fe48f45b2877e85522eee84029b12d3e0ae5b91" Oct 02 10:07:37 crc kubenswrapper[4934]: I1002 10:07:37.104497 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/edb76b11-9129-495d-b44d-e998c3e8dceb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"edb76b11-9129-495d-b44d-e998c3e8dceb\") " pod="openstack/ceilometer-0" Oct 02 10:07:37 crc kubenswrapper[4934]: I1002 10:07:37.104613 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edb76b11-9129-495d-b44d-e998c3e8dceb-scripts\") pod \"ceilometer-0\" (UID: \"edb76b11-9129-495d-b44d-e998c3e8dceb\") " pod="openstack/ceilometer-0" Oct 02 10:07:37 crc kubenswrapper[4934]: I1002 10:07:37.104645 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edb76b11-9129-495d-b44d-e998c3e8dceb-run-httpd\") pod \"ceilometer-0\" (UID: \"edb76b11-9129-495d-b44d-e998c3e8dceb\") " pod="openstack/ceilometer-0" Oct 02 10:07:37 crc kubenswrapper[4934]: I1002 10:07:37.104731 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edb76b11-9129-495d-b44d-e998c3e8dceb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"edb76b11-9129-495d-b44d-e998c3e8dceb\") " pod="openstack/ceilometer-0" Oct 02 10:07:37 crc kubenswrapper[4934]: I1002 10:07:37.104759 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edb76b11-9129-495d-b44d-e998c3e8dceb-config-data\") pod \"ceilometer-0\" (UID: \"edb76b11-9129-495d-b44d-e998c3e8dceb\") " pod="openstack/ceilometer-0" Oct 02 10:07:37 crc kubenswrapper[4934]: I1002 10:07:37.104782 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79x2b\" (UniqueName: \"kubernetes.io/projected/edb76b11-9129-495d-b44d-e998c3e8dceb-kube-api-access-79x2b\") pod \"ceilometer-0\" (UID: \"edb76b11-9129-495d-b44d-e998c3e8dceb\") " pod="openstack/ceilometer-0" Oct 02 10:07:37 crc kubenswrapper[4934]: I1002 10:07:37.104805 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edb76b11-9129-495d-b44d-e998c3e8dceb-log-httpd\") pod \"ceilometer-0\" (UID: \"edb76b11-9129-495d-b44d-e998c3e8dceb\") " pod="openstack/ceilometer-0" Oct 02 10:07:37 crc kubenswrapper[4934]: I1002 10:07:37.206336 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edb76b11-9129-495d-b44d-e998c3e8dceb-scripts\") pod \"ceilometer-0\" (UID: \"edb76b11-9129-495d-b44d-e998c3e8dceb\") " pod="openstack/ceilometer-0" Oct 02 10:07:37 crc kubenswrapper[4934]: I1002 10:07:37.206396 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edb76b11-9129-495d-b44d-e998c3e8dceb-run-httpd\") pod \"ceilometer-0\" (UID: \"edb76b11-9129-495d-b44d-e998c3e8dceb\") " pod="openstack/ceilometer-0" Oct 02 10:07:37 crc kubenswrapper[4934]: I1002 10:07:37.206460 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edb76b11-9129-495d-b44d-e998c3e8dceb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"edb76b11-9129-495d-b44d-e998c3e8dceb\") " pod="openstack/ceilometer-0" Oct 02 10:07:37 crc kubenswrapper[4934]: I1002 10:07:37.206489 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edb76b11-9129-495d-b44d-e998c3e8dceb-config-data\") pod \"ceilometer-0\" (UID: \"edb76b11-9129-495d-b44d-e998c3e8dceb\") " pod="openstack/ceilometer-0" Oct 02 10:07:37 crc kubenswrapper[4934]: I1002 10:07:37.206508 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79x2b\" (UniqueName: \"kubernetes.io/projected/edb76b11-9129-495d-b44d-e998c3e8dceb-kube-api-access-79x2b\") pod \"ceilometer-0\" (UID: \"edb76b11-9129-495d-b44d-e998c3e8dceb\") " pod="openstack/ceilometer-0" Oct 02 10:07:37 crc kubenswrapper[4934]: I1002 10:07:37.206524 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edb76b11-9129-495d-b44d-e998c3e8dceb-log-httpd\") pod \"ceilometer-0\" (UID: \"edb76b11-9129-495d-b44d-e998c3e8dceb\") " pod="openstack/ceilometer-0" Oct 02 10:07:37 crc kubenswrapper[4934]: I1002 10:07:37.206550 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/edb76b11-9129-495d-b44d-e998c3e8dceb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"edb76b11-9129-495d-b44d-e998c3e8dceb\") " pod="openstack/ceilometer-0" Oct 02 10:07:37 crc kubenswrapper[4934]: I1002 10:07:37.209940 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edb76b11-9129-495d-b44d-e998c3e8dceb-run-httpd\") pod \"ceilometer-0\" (UID: \"edb76b11-9129-495d-b44d-e998c3e8dceb\") " pod="openstack/ceilometer-0" Oct 02 10:07:37 crc kubenswrapper[4934]: I1002 10:07:37.210008 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edb76b11-9129-495d-b44d-e998c3e8dceb-log-httpd\") pod \"ceilometer-0\" (UID: \"edb76b11-9129-495d-b44d-e998c3e8dceb\") " pod="openstack/ceilometer-0" Oct 02 10:07:37 crc kubenswrapper[4934]: I1002 10:07:37.210215 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/edb76b11-9129-495d-b44d-e998c3e8dceb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"edb76b11-9129-495d-b44d-e998c3e8dceb\") " pod="openstack/ceilometer-0" Oct 02 10:07:37 crc kubenswrapper[4934]: I1002 10:07:37.213113 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edb76b11-9129-495d-b44d-e998c3e8dceb-scripts\") pod \"ceilometer-0\" (UID: \"edb76b11-9129-495d-b44d-e998c3e8dceb\") " pod="openstack/ceilometer-0" Oct 02 10:07:37 crc kubenswrapper[4934]: I1002 10:07:37.213439 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edb76b11-9129-495d-b44d-e998c3e8dceb-config-data\") pod \"ceilometer-0\" (UID: \"edb76b11-9129-495d-b44d-e998c3e8dceb\") " pod="openstack/ceilometer-0" Oct 02 10:07:37 crc kubenswrapper[4934]: I1002 10:07:37.214040 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edb76b11-9129-495d-b44d-e998c3e8dceb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"edb76b11-9129-495d-b44d-e998c3e8dceb\") " pod="openstack/ceilometer-0" Oct 02 10:07:37 crc kubenswrapper[4934]: I1002 10:07:37.232115 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79x2b\" (UniqueName: \"kubernetes.io/projected/edb76b11-9129-495d-b44d-e998c3e8dceb-kube-api-access-79x2b\") pod \"ceilometer-0\" (UID: \"edb76b11-9129-495d-b44d-e998c3e8dceb\") " pod="openstack/ceilometer-0" Oct 02 10:07:37 crc kubenswrapper[4934]: I1002 10:07:37.298081 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:07:37 crc kubenswrapper[4934]: I1002 10:07:37.734748 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:07:37 crc kubenswrapper[4934]: W1002 10:07:37.739165 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podedb76b11_9129_495d_b44d_e998c3e8dceb.slice/crio-1997b7563b3c0b444d269d563331d094f5e6ccb2898a0f7c1adccc61176a2e00 WatchSource:0}: Error finding container 1997b7563b3c0b444d269d563331d094f5e6ccb2898a0f7c1adccc61176a2e00: Status 404 returned error can't find the container with id 1997b7563b3c0b444d269d563331d094f5e6ccb2898a0f7c1adccc61176a2e00 Oct 02 10:07:37 crc kubenswrapper[4934]: I1002 10:07:37.900202 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"edb76b11-9129-495d-b44d-e998c3e8dceb","Type":"ContainerStarted","Data":"1997b7563b3c0b444d269d563331d094f5e6ccb2898a0f7c1adccc61176a2e00"} Oct 02 10:07:38 crc kubenswrapper[4934]: I1002 10:07:38.439719 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:07:38 crc kubenswrapper[4934]: I1002 10:07:38.439791 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:07:38 crc kubenswrapper[4934]: I1002 10:07:38.910203 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"edb76b11-9129-495d-b44d-e998c3e8dceb","Type":"ContainerStarted","Data":"3e597152c8970caf6df4198a19c0fb1ca0108eab7f0f4e07d15ecd0797056ed0"} Oct 02 10:07:38 crc kubenswrapper[4934]: I1002 10:07:38.924440 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a01b1d99-451c-4aec-b261-53232f25e582" path="/var/lib/kubelet/pods/a01b1d99-451c-4aec-b261-53232f25e582/volumes" Oct 02 10:07:39 crc kubenswrapper[4934]: I1002 10:07:39.922589 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"edb76b11-9129-495d-b44d-e998c3e8dceb","Type":"ContainerStarted","Data":"cee5b6bd1bcc9c2981f3668a02ad37893a0ce6cb09bbaf102043a1606140b34f"} Oct 02 10:07:40 crc kubenswrapper[4934]: I1002 10:07:40.930020 4934 generic.go:334] "Generic (PLEG): container finished" podID="157aa366-544e-42ba-bd81-1649575c976f" containerID="9d35a7ad52afadce99f8b4bb3bb5a4238db85b29282f3f5f2fa42c6bd9b238ef" exitCode=0 Oct 02 10:07:40 crc kubenswrapper[4934]: I1002 10:07:40.930068 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-klshd" event={"ID":"157aa366-544e-42ba-bd81-1649575c976f","Type":"ContainerDied","Data":"9d35a7ad52afadce99f8b4bb3bb5a4238db85b29282f3f5f2fa42c6bd9b238ef"} Oct 02 10:07:40 crc kubenswrapper[4934]: I1002 10:07:40.932351 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"edb76b11-9129-495d-b44d-e998c3e8dceb","Type":"ContainerStarted","Data":"3ae22becac0e4812e95a08cb7d2bbcf1e151ec55f7db552c2af6ec857bd58070"} Oct 02 10:07:41 crc kubenswrapper[4934]: I1002 10:07:41.947107 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"edb76b11-9129-495d-b44d-e998c3e8dceb","Type":"ContainerStarted","Data":"07a3ebc27c86cc0989e83071c71d3c33a754a7f416d67889d187120a9dc11c03"} Oct 02 10:07:41 crc kubenswrapper[4934]: I1002 10:07:41.947342 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 10:07:42 crc kubenswrapper[4934]: I1002 10:07:42.293519 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-klshd" Oct 02 10:07:42 crc kubenswrapper[4934]: I1002 10:07:42.312436 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.8349089530000002 podStartE2EDuration="6.312411892s" podCreationTimestamp="2025-10-02 10:07:36 +0000 UTC" firstStartedPulling="2025-10-02 10:07:37.74156792 +0000 UTC m=+1129.494209452" lastFinishedPulling="2025-10-02 10:07:41.219070869 +0000 UTC m=+1132.971712391" observedRunningTime="2025-10-02 10:07:41.977915168 +0000 UTC m=+1133.730556680" watchObservedRunningTime="2025-10-02 10:07:42.312411892 +0000 UTC m=+1134.065053414" Oct 02 10:07:42 crc kubenswrapper[4934]: I1002 10:07:42.419124 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jp9h\" (UniqueName: \"kubernetes.io/projected/157aa366-544e-42ba-bd81-1649575c976f-kube-api-access-4jp9h\") pod \"157aa366-544e-42ba-bd81-1649575c976f\" (UID: \"157aa366-544e-42ba-bd81-1649575c976f\") " Oct 02 10:07:42 crc kubenswrapper[4934]: I1002 10:07:42.419320 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/157aa366-544e-42ba-bd81-1649575c976f-combined-ca-bundle\") pod \"157aa366-544e-42ba-bd81-1649575c976f\" (UID: \"157aa366-544e-42ba-bd81-1649575c976f\") " Oct 02 10:07:42 crc kubenswrapper[4934]: I1002 10:07:42.419366 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/157aa366-544e-42ba-bd81-1649575c976f-scripts\") pod \"157aa366-544e-42ba-bd81-1649575c976f\" (UID: \"157aa366-544e-42ba-bd81-1649575c976f\") " Oct 02 10:07:42 crc kubenswrapper[4934]: I1002 10:07:42.419424 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/157aa366-544e-42ba-bd81-1649575c976f-logs\") pod \"157aa366-544e-42ba-bd81-1649575c976f\" (UID: \"157aa366-544e-42ba-bd81-1649575c976f\") " Oct 02 10:07:42 crc kubenswrapper[4934]: I1002 10:07:42.419457 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/157aa366-544e-42ba-bd81-1649575c976f-config-data\") pod \"157aa366-544e-42ba-bd81-1649575c976f\" (UID: \"157aa366-544e-42ba-bd81-1649575c976f\") " Oct 02 10:07:42 crc kubenswrapper[4934]: I1002 10:07:42.428904 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/157aa366-544e-42ba-bd81-1649575c976f-logs" (OuterVolumeSpecName: "logs") pod "157aa366-544e-42ba-bd81-1649575c976f" (UID: "157aa366-544e-42ba-bd81-1649575c976f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:07:42 crc kubenswrapper[4934]: I1002 10:07:42.432543 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/157aa366-544e-42ba-bd81-1649575c976f-scripts" (OuterVolumeSpecName: "scripts") pod "157aa366-544e-42ba-bd81-1649575c976f" (UID: "157aa366-544e-42ba-bd81-1649575c976f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:07:42 crc kubenswrapper[4934]: I1002 10:07:42.432639 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/157aa366-544e-42ba-bd81-1649575c976f-kube-api-access-4jp9h" (OuterVolumeSpecName: "kube-api-access-4jp9h") pod "157aa366-544e-42ba-bd81-1649575c976f" (UID: "157aa366-544e-42ba-bd81-1649575c976f"). InnerVolumeSpecName "kube-api-access-4jp9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:07:42 crc kubenswrapper[4934]: I1002 10:07:42.464969 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/157aa366-544e-42ba-bd81-1649575c976f-config-data" (OuterVolumeSpecName: "config-data") pod "157aa366-544e-42ba-bd81-1649575c976f" (UID: "157aa366-544e-42ba-bd81-1649575c976f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:07:42 crc kubenswrapper[4934]: I1002 10:07:42.484881 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/157aa366-544e-42ba-bd81-1649575c976f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "157aa366-544e-42ba-bd81-1649575c976f" (UID: "157aa366-544e-42ba-bd81-1649575c976f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:07:42 crc kubenswrapper[4934]: I1002 10:07:42.521355 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/157aa366-544e-42ba-bd81-1649575c976f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:42 crc kubenswrapper[4934]: I1002 10:07:42.521387 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/157aa366-544e-42ba-bd81-1649575c976f-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:42 crc kubenswrapper[4934]: I1002 10:07:42.521397 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/157aa366-544e-42ba-bd81-1649575c976f-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:42 crc kubenswrapper[4934]: I1002 10:07:42.521405 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/157aa366-544e-42ba-bd81-1649575c976f-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:42 crc kubenswrapper[4934]: I1002 10:07:42.521413 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jp9h\" (UniqueName: \"kubernetes.io/projected/157aa366-544e-42ba-bd81-1649575c976f-kube-api-access-4jp9h\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:42 crc kubenswrapper[4934]: I1002 10:07:42.958445 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-klshd" Oct 02 10:07:42 crc kubenswrapper[4934]: I1002 10:07:42.959098 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-klshd" event={"ID":"157aa366-544e-42ba-bd81-1649575c976f","Type":"ContainerDied","Data":"4501e43d704d24923d3a1889882ed517cc126f212d5c40ef28b2fbf054abc713"} Oct 02 10:07:42 crc kubenswrapper[4934]: I1002 10:07:42.959123 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4501e43d704d24923d3a1889882ed517cc126f212d5c40ef28b2fbf054abc713" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.150561 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-649ccd9666-dh5rs"] Oct 02 10:07:43 crc kubenswrapper[4934]: E1002 10:07:43.151052 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="157aa366-544e-42ba-bd81-1649575c976f" containerName="placement-db-sync" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.151075 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="157aa366-544e-42ba-bd81-1649575c976f" containerName="placement-db-sync" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.151309 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="157aa366-544e-42ba-bd81-1649575c976f" containerName="placement-db-sync" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.152296 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.155120 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-t9vdb" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.155516 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.155739 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.155945 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.156041 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.164772 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-649ccd9666-dh5rs"] Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.233489 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-config-data\") pod \"placement-649ccd9666-dh5rs\" (UID: \"ecc76715-8b05-4529-bd7b-289d7f32eff5\") " pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.233569 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-public-tls-certs\") pod \"placement-649ccd9666-dh5rs\" (UID: \"ecc76715-8b05-4529-bd7b-289d7f32eff5\") " pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.233614 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-internal-tls-certs\") pod \"placement-649ccd9666-dh5rs\" (UID: \"ecc76715-8b05-4529-bd7b-289d7f32eff5\") " pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.233723 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecc76715-8b05-4529-bd7b-289d7f32eff5-logs\") pod \"placement-649ccd9666-dh5rs\" (UID: \"ecc76715-8b05-4529-bd7b-289d7f32eff5\") " pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.233765 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-scripts\") pod \"placement-649ccd9666-dh5rs\" (UID: \"ecc76715-8b05-4529-bd7b-289d7f32eff5\") " pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.233791 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vk7k\" (UniqueName: \"kubernetes.io/projected/ecc76715-8b05-4529-bd7b-289d7f32eff5-kube-api-access-2vk7k\") pod \"placement-649ccd9666-dh5rs\" (UID: \"ecc76715-8b05-4529-bd7b-289d7f32eff5\") " pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.233824 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-combined-ca-bundle\") pod \"placement-649ccd9666-dh5rs\" (UID: \"ecc76715-8b05-4529-bd7b-289d7f32eff5\") " pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.335453 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-public-tls-certs\") pod \"placement-649ccd9666-dh5rs\" (UID: \"ecc76715-8b05-4529-bd7b-289d7f32eff5\") " pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.335508 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-internal-tls-certs\") pod \"placement-649ccd9666-dh5rs\" (UID: \"ecc76715-8b05-4529-bd7b-289d7f32eff5\") " pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.335624 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecc76715-8b05-4529-bd7b-289d7f32eff5-logs\") pod \"placement-649ccd9666-dh5rs\" (UID: \"ecc76715-8b05-4529-bd7b-289d7f32eff5\") " pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.335655 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-scripts\") pod \"placement-649ccd9666-dh5rs\" (UID: \"ecc76715-8b05-4529-bd7b-289d7f32eff5\") " pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.335679 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vk7k\" (UniqueName: \"kubernetes.io/projected/ecc76715-8b05-4529-bd7b-289d7f32eff5-kube-api-access-2vk7k\") pod \"placement-649ccd9666-dh5rs\" (UID: \"ecc76715-8b05-4529-bd7b-289d7f32eff5\") " pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.335717 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-combined-ca-bundle\") pod \"placement-649ccd9666-dh5rs\" (UID: \"ecc76715-8b05-4529-bd7b-289d7f32eff5\") " pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.335756 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-config-data\") pod \"placement-649ccd9666-dh5rs\" (UID: \"ecc76715-8b05-4529-bd7b-289d7f32eff5\") " pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.336665 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecc76715-8b05-4529-bd7b-289d7f32eff5-logs\") pod \"placement-649ccd9666-dh5rs\" (UID: \"ecc76715-8b05-4529-bd7b-289d7f32eff5\") " pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.339495 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-combined-ca-bundle\") pod \"placement-649ccd9666-dh5rs\" (UID: \"ecc76715-8b05-4529-bd7b-289d7f32eff5\") " pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.340396 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-config-data\") pod \"placement-649ccd9666-dh5rs\" (UID: \"ecc76715-8b05-4529-bd7b-289d7f32eff5\") " pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.340703 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-public-tls-certs\") pod \"placement-649ccd9666-dh5rs\" (UID: \"ecc76715-8b05-4529-bd7b-289d7f32eff5\") " pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.341227 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-internal-tls-certs\") pod \"placement-649ccd9666-dh5rs\" (UID: \"ecc76715-8b05-4529-bd7b-289d7f32eff5\") " pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.347103 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-scripts\") pod \"placement-649ccd9666-dh5rs\" (UID: \"ecc76715-8b05-4529-bd7b-289d7f32eff5\") " pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.368098 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vk7k\" (UniqueName: \"kubernetes.io/projected/ecc76715-8b05-4529-bd7b-289d7f32eff5-kube-api-access-2vk7k\") pod \"placement-649ccd9666-dh5rs\" (UID: \"ecc76715-8b05-4529-bd7b-289d7f32eff5\") " pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.474221 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.944821 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-649ccd9666-dh5rs"] Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.971965 4934 generic.go:334] "Generic (PLEG): container finished" podID="379bc476-7f65-408d-929a-8badb1e4dfdc" containerID="d4c19f695157dd25d239073bba217146b7e1c30d74c166a1697941fd5eab8a53" exitCode=0 Oct 02 10:07:43 crc kubenswrapper[4934]: I1002 10:07:43.972021 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-rqm5f" event={"ID":"379bc476-7f65-408d-929a-8badb1e4dfdc","Type":"ContainerDied","Data":"d4c19f695157dd25d239073bba217146b7e1c30d74c166a1697941fd5eab8a53"} Oct 02 10:07:44 crc kubenswrapper[4934]: I1002 10:07:44.982695 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-649ccd9666-dh5rs" event={"ID":"ecc76715-8b05-4529-bd7b-289d7f32eff5","Type":"ContainerStarted","Data":"8710e544f71f9ca7366446c35d7eb3080fa07ae1e0f59d3a870062d5e14b9eb6"} Oct 02 10:07:44 crc kubenswrapper[4934]: I1002 10:07:44.983057 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-649ccd9666-dh5rs" event={"ID":"ecc76715-8b05-4529-bd7b-289d7f32eff5","Type":"ContainerStarted","Data":"4d19c2050c253cd17d4e337c977a53e89370af2b71d1d6898f8b9f9d88a2f2b2"} Oct 02 10:07:44 crc kubenswrapper[4934]: I1002 10:07:44.983071 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-649ccd9666-dh5rs" event={"ID":"ecc76715-8b05-4529-bd7b-289d7f32eff5","Type":"ContainerStarted","Data":"e4d07638d617ac5b5a4bf2034509973c9924de81e2c70de5faa9fb5370268e3c"} Oct 02 10:07:45 crc kubenswrapper[4934]: I1002 10:07:45.006274 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-649ccd9666-dh5rs" podStartSLOduration=2.006252605 podStartE2EDuration="2.006252605s" podCreationTimestamp="2025-10-02 10:07:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:07:45.001791198 +0000 UTC m=+1136.754432740" watchObservedRunningTime="2025-10-02 10:07:45.006252605 +0000 UTC m=+1136.758894137" Oct 02 10:07:45 crc kubenswrapper[4934]: I1002 10:07:45.329606 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-rqm5f" Oct 02 10:07:45 crc kubenswrapper[4934]: I1002 10:07:45.482652 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/379bc476-7f65-408d-929a-8badb1e4dfdc-db-sync-config-data\") pod \"379bc476-7f65-408d-929a-8badb1e4dfdc\" (UID: \"379bc476-7f65-408d-929a-8badb1e4dfdc\") " Oct 02 10:07:45 crc kubenswrapper[4934]: I1002 10:07:45.483104 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmzvx\" (UniqueName: \"kubernetes.io/projected/379bc476-7f65-408d-929a-8badb1e4dfdc-kube-api-access-bmzvx\") pod \"379bc476-7f65-408d-929a-8badb1e4dfdc\" (UID: \"379bc476-7f65-408d-929a-8badb1e4dfdc\") " Oct 02 10:07:45 crc kubenswrapper[4934]: I1002 10:07:45.483219 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/379bc476-7f65-408d-929a-8badb1e4dfdc-combined-ca-bundle\") pod \"379bc476-7f65-408d-929a-8badb1e4dfdc\" (UID: \"379bc476-7f65-408d-929a-8badb1e4dfdc\") " Oct 02 10:07:45 crc kubenswrapper[4934]: I1002 10:07:45.489624 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/379bc476-7f65-408d-929a-8badb1e4dfdc-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "379bc476-7f65-408d-929a-8badb1e4dfdc" (UID: "379bc476-7f65-408d-929a-8badb1e4dfdc"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:07:45 crc kubenswrapper[4934]: I1002 10:07:45.497532 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/379bc476-7f65-408d-929a-8badb1e4dfdc-kube-api-access-bmzvx" (OuterVolumeSpecName: "kube-api-access-bmzvx") pod "379bc476-7f65-408d-929a-8badb1e4dfdc" (UID: "379bc476-7f65-408d-929a-8badb1e4dfdc"). InnerVolumeSpecName "kube-api-access-bmzvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:07:45 crc kubenswrapper[4934]: I1002 10:07:45.529697 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/379bc476-7f65-408d-929a-8badb1e4dfdc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "379bc476-7f65-408d-929a-8badb1e4dfdc" (UID: "379bc476-7f65-408d-929a-8badb1e4dfdc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:07:45 crc kubenswrapper[4934]: I1002 10:07:45.585143 4934 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/379bc476-7f65-408d-929a-8badb1e4dfdc-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:45 crc kubenswrapper[4934]: I1002 10:07:45.585207 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmzvx\" (UniqueName: \"kubernetes.io/projected/379bc476-7f65-408d-929a-8badb1e4dfdc-kube-api-access-bmzvx\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:45 crc kubenswrapper[4934]: I1002 10:07:45.585234 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/379bc476-7f65-408d-929a-8badb1e4dfdc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:45 crc kubenswrapper[4934]: I1002 10:07:45.991431 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-rqm5f" event={"ID":"379bc476-7f65-408d-929a-8badb1e4dfdc","Type":"ContainerDied","Data":"776b050a8b5e2fb2457cecd8cb868628b610293ebb85322da88f7180123b6a94"} Oct 02 10:07:45 crc kubenswrapper[4934]: I1002 10:07:45.991484 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="776b050a8b5e2fb2457cecd8cb868628b610293ebb85322da88f7180123b6a94" Oct 02 10:07:45 crc kubenswrapper[4934]: I1002 10:07:45.991627 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-rqm5f" Oct 02 10:07:45 crc kubenswrapper[4934]: I1002 10:07:45.991646 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:07:45 crc kubenswrapper[4934]: I1002 10:07:45.992325 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.295166 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-6bf48f9d67-g95nd"] Oct 02 10:07:46 crc kubenswrapper[4934]: E1002 10:07:46.295884 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="379bc476-7f65-408d-929a-8badb1e4dfdc" containerName="barbican-db-sync" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.295908 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="379bc476-7f65-408d-929a-8badb1e4dfdc" containerName="barbican-db-sync" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.296120 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="379bc476-7f65-408d-929a-8badb1e4dfdc" containerName="barbican-db-sync" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.297774 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6bf48f9d67-g95nd" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.299698 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.299838 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-5l4m9" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.301497 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.307693 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6bf48f9d67-g95nd"] Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.319169 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-b97fc8746-vlx9z"] Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.327972 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-b97fc8746-vlx9z" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.330549 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.364924 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-b97fc8746-vlx9z"] Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.382872 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-867d6b4989-5mkmw"] Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.384438 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.398200 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/547c6c6d-0f01-4feb-acc6-2e3045407b64-config-data-custom\") pod \"barbican-worker-6bf48f9d67-g95nd\" (UID: \"547c6c6d-0f01-4feb-acc6-2e3045407b64\") " pod="openstack/barbican-worker-6bf48f9d67-g95nd" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.398249 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/547c6c6d-0f01-4feb-acc6-2e3045407b64-logs\") pod \"barbican-worker-6bf48f9d67-g95nd\" (UID: \"547c6c6d-0f01-4feb-acc6-2e3045407b64\") " pod="openstack/barbican-worker-6bf48f9d67-g95nd" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.398291 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/547c6c6d-0f01-4feb-acc6-2e3045407b64-config-data\") pod \"barbican-worker-6bf48f9d67-g95nd\" (UID: \"547c6c6d-0f01-4feb-acc6-2e3045407b64\") " pod="openstack/barbican-worker-6bf48f9d67-g95nd" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.398338 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxzdw\" (UniqueName: \"kubernetes.io/projected/547c6c6d-0f01-4feb-acc6-2e3045407b64-kube-api-access-vxzdw\") pod \"barbican-worker-6bf48f9d67-g95nd\" (UID: \"547c6c6d-0f01-4feb-acc6-2e3045407b64\") " pod="openstack/barbican-worker-6bf48f9d67-g95nd" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.398367 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/547c6c6d-0f01-4feb-acc6-2e3045407b64-combined-ca-bundle\") pod \"barbican-worker-6bf48f9d67-g95nd\" (UID: \"547c6c6d-0f01-4feb-acc6-2e3045407b64\") " pod="openstack/barbican-worker-6bf48f9d67-g95nd" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.415495 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-867d6b4989-5mkmw"] Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.499541 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwzdf\" (UniqueName: \"kubernetes.io/projected/19a55666-374f-430a-a80b-f61538cbfb6b-kube-api-access-qwzdf\") pod \"dnsmasq-dns-867d6b4989-5mkmw\" (UID: \"19a55666-374f-430a-a80b-f61538cbfb6b\") " pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.499601 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-ovsdbserver-sb\") pod \"dnsmasq-dns-867d6b4989-5mkmw\" (UID: \"19a55666-374f-430a-a80b-f61538cbfb6b\") " pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.499627 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-dns-swift-storage-0\") pod \"dnsmasq-dns-867d6b4989-5mkmw\" (UID: \"19a55666-374f-430a-a80b-f61538cbfb6b\") " pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.499657 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/547c6c6d-0f01-4feb-acc6-2e3045407b64-config-data-custom\") pod \"barbican-worker-6bf48f9d67-g95nd\" (UID: \"547c6c6d-0f01-4feb-acc6-2e3045407b64\") " pod="openstack/barbican-worker-6bf48f9d67-g95nd" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.499682 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-dns-svc\") pod \"dnsmasq-dns-867d6b4989-5mkmw\" (UID: \"19a55666-374f-430a-a80b-f61538cbfb6b\") " pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.499736 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/547c6c6d-0f01-4feb-acc6-2e3045407b64-logs\") pod \"barbican-worker-6bf48f9d67-g95nd\" (UID: \"547c6c6d-0f01-4feb-acc6-2e3045407b64\") " pod="openstack/barbican-worker-6bf48f9d67-g95nd" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.499776 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/547c6c6d-0f01-4feb-acc6-2e3045407b64-config-data\") pod \"barbican-worker-6bf48f9d67-g95nd\" (UID: \"547c6c6d-0f01-4feb-acc6-2e3045407b64\") " pod="openstack/barbican-worker-6bf48f9d67-g95nd" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.499814 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d15b0450-e100-4ee7-ad62-88601eedc4f2-logs\") pod \"barbican-keystone-listener-b97fc8746-vlx9z\" (UID: \"d15b0450-e100-4ee7-ad62-88601eedc4f2\") " pod="openstack/barbican-keystone-listener-b97fc8746-vlx9z" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.499839 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d15b0450-e100-4ee7-ad62-88601eedc4f2-config-data\") pod \"barbican-keystone-listener-b97fc8746-vlx9z\" (UID: \"d15b0450-e100-4ee7-ad62-88601eedc4f2\") " pod="openstack/barbican-keystone-listener-b97fc8746-vlx9z" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.499858 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtknw\" (UniqueName: \"kubernetes.io/projected/d15b0450-e100-4ee7-ad62-88601eedc4f2-kube-api-access-vtknw\") pod \"barbican-keystone-listener-b97fc8746-vlx9z\" (UID: \"d15b0450-e100-4ee7-ad62-88601eedc4f2\") " pod="openstack/barbican-keystone-listener-b97fc8746-vlx9z" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.499875 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxzdw\" (UniqueName: \"kubernetes.io/projected/547c6c6d-0f01-4feb-acc6-2e3045407b64-kube-api-access-vxzdw\") pod \"barbican-worker-6bf48f9d67-g95nd\" (UID: \"547c6c6d-0f01-4feb-acc6-2e3045407b64\") " pod="openstack/barbican-worker-6bf48f9d67-g95nd" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.499902 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d15b0450-e100-4ee7-ad62-88601eedc4f2-config-data-custom\") pod \"barbican-keystone-listener-b97fc8746-vlx9z\" (UID: \"d15b0450-e100-4ee7-ad62-88601eedc4f2\") " pod="openstack/barbican-keystone-listener-b97fc8746-vlx9z" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.499919 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-config\") pod \"dnsmasq-dns-867d6b4989-5mkmw\" (UID: \"19a55666-374f-430a-a80b-f61538cbfb6b\") " pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.499935 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-ovsdbserver-nb\") pod \"dnsmasq-dns-867d6b4989-5mkmw\" (UID: \"19a55666-374f-430a-a80b-f61538cbfb6b\") " pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.499954 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/547c6c6d-0f01-4feb-acc6-2e3045407b64-combined-ca-bundle\") pod \"barbican-worker-6bf48f9d67-g95nd\" (UID: \"547c6c6d-0f01-4feb-acc6-2e3045407b64\") " pod="openstack/barbican-worker-6bf48f9d67-g95nd" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.499973 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d15b0450-e100-4ee7-ad62-88601eedc4f2-combined-ca-bundle\") pod \"barbican-keystone-listener-b97fc8746-vlx9z\" (UID: \"d15b0450-e100-4ee7-ad62-88601eedc4f2\") " pod="openstack/barbican-keystone-listener-b97fc8746-vlx9z" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.500479 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/547c6c6d-0f01-4feb-acc6-2e3045407b64-logs\") pod \"barbican-worker-6bf48f9d67-g95nd\" (UID: \"547c6c6d-0f01-4feb-acc6-2e3045407b64\") " pod="openstack/barbican-worker-6bf48f9d67-g95nd" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.513701 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/547c6c6d-0f01-4feb-acc6-2e3045407b64-combined-ca-bundle\") pod \"barbican-worker-6bf48f9d67-g95nd\" (UID: \"547c6c6d-0f01-4feb-acc6-2e3045407b64\") " pod="openstack/barbican-worker-6bf48f9d67-g95nd" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.514628 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/547c6c6d-0f01-4feb-acc6-2e3045407b64-config-data\") pod \"barbican-worker-6bf48f9d67-g95nd\" (UID: \"547c6c6d-0f01-4feb-acc6-2e3045407b64\") " pod="openstack/barbican-worker-6bf48f9d67-g95nd" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.523151 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxzdw\" (UniqueName: \"kubernetes.io/projected/547c6c6d-0f01-4feb-acc6-2e3045407b64-kube-api-access-vxzdw\") pod \"barbican-worker-6bf48f9d67-g95nd\" (UID: \"547c6c6d-0f01-4feb-acc6-2e3045407b64\") " pod="openstack/barbican-worker-6bf48f9d67-g95nd" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.535633 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/547c6c6d-0f01-4feb-acc6-2e3045407b64-config-data-custom\") pod \"barbican-worker-6bf48f9d67-g95nd\" (UID: \"547c6c6d-0f01-4feb-acc6-2e3045407b64\") " pod="openstack/barbican-worker-6bf48f9d67-g95nd" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.563266 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-57c45ff45d-ghdvr"] Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.564646 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57c45ff45d-ghdvr" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.569101 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.596999 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-57c45ff45d-ghdvr"] Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.601075 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d15b0450-e100-4ee7-ad62-88601eedc4f2-logs\") pod \"barbican-keystone-listener-b97fc8746-vlx9z\" (UID: \"d15b0450-e100-4ee7-ad62-88601eedc4f2\") " pod="openstack/barbican-keystone-listener-b97fc8746-vlx9z" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.601119 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d15b0450-e100-4ee7-ad62-88601eedc4f2-config-data\") pod \"barbican-keystone-listener-b97fc8746-vlx9z\" (UID: \"d15b0450-e100-4ee7-ad62-88601eedc4f2\") " pod="openstack/barbican-keystone-listener-b97fc8746-vlx9z" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.601145 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtknw\" (UniqueName: \"kubernetes.io/projected/d15b0450-e100-4ee7-ad62-88601eedc4f2-kube-api-access-vtknw\") pod \"barbican-keystone-listener-b97fc8746-vlx9z\" (UID: \"d15b0450-e100-4ee7-ad62-88601eedc4f2\") " pod="openstack/barbican-keystone-listener-b97fc8746-vlx9z" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.601174 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d15b0450-e100-4ee7-ad62-88601eedc4f2-config-data-custom\") pod \"barbican-keystone-listener-b97fc8746-vlx9z\" (UID: \"d15b0450-e100-4ee7-ad62-88601eedc4f2\") " pod="openstack/barbican-keystone-listener-b97fc8746-vlx9z" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.601190 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-config\") pod \"dnsmasq-dns-867d6b4989-5mkmw\" (UID: \"19a55666-374f-430a-a80b-f61538cbfb6b\") " pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.601208 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-ovsdbserver-nb\") pod \"dnsmasq-dns-867d6b4989-5mkmw\" (UID: \"19a55666-374f-430a-a80b-f61538cbfb6b\") " pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.601230 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d15b0450-e100-4ee7-ad62-88601eedc4f2-combined-ca-bundle\") pod \"barbican-keystone-listener-b97fc8746-vlx9z\" (UID: \"d15b0450-e100-4ee7-ad62-88601eedc4f2\") " pod="openstack/barbican-keystone-listener-b97fc8746-vlx9z" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.601273 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwzdf\" (UniqueName: \"kubernetes.io/projected/19a55666-374f-430a-a80b-f61538cbfb6b-kube-api-access-qwzdf\") pod \"dnsmasq-dns-867d6b4989-5mkmw\" (UID: \"19a55666-374f-430a-a80b-f61538cbfb6b\") " pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.601300 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-ovsdbserver-sb\") pod \"dnsmasq-dns-867d6b4989-5mkmw\" (UID: \"19a55666-374f-430a-a80b-f61538cbfb6b\") " pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.601337 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-dns-swift-storage-0\") pod \"dnsmasq-dns-867d6b4989-5mkmw\" (UID: \"19a55666-374f-430a-a80b-f61538cbfb6b\") " pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.601366 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-dns-svc\") pod \"dnsmasq-dns-867d6b4989-5mkmw\" (UID: \"19a55666-374f-430a-a80b-f61538cbfb6b\") " pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.602758 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-dns-svc\") pod \"dnsmasq-dns-867d6b4989-5mkmw\" (UID: \"19a55666-374f-430a-a80b-f61538cbfb6b\") " pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.602913 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-config\") pod \"dnsmasq-dns-867d6b4989-5mkmw\" (UID: \"19a55666-374f-430a-a80b-f61538cbfb6b\") " pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.603178 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d15b0450-e100-4ee7-ad62-88601eedc4f2-logs\") pod \"barbican-keystone-listener-b97fc8746-vlx9z\" (UID: \"d15b0450-e100-4ee7-ad62-88601eedc4f2\") " pod="openstack/barbican-keystone-listener-b97fc8746-vlx9z" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.603288 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-ovsdbserver-nb\") pod \"dnsmasq-dns-867d6b4989-5mkmw\" (UID: \"19a55666-374f-430a-a80b-f61538cbfb6b\") " pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.605149 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-ovsdbserver-sb\") pod \"dnsmasq-dns-867d6b4989-5mkmw\" (UID: \"19a55666-374f-430a-a80b-f61538cbfb6b\") " pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.605987 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-dns-swift-storage-0\") pod \"dnsmasq-dns-867d6b4989-5mkmw\" (UID: \"19a55666-374f-430a-a80b-f61538cbfb6b\") " pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.607505 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d15b0450-e100-4ee7-ad62-88601eedc4f2-combined-ca-bundle\") pod \"barbican-keystone-listener-b97fc8746-vlx9z\" (UID: \"d15b0450-e100-4ee7-ad62-88601eedc4f2\") " pod="openstack/barbican-keystone-listener-b97fc8746-vlx9z" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.610078 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d15b0450-e100-4ee7-ad62-88601eedc4f2-config-data-custom\") pod \"barbican-keystone-listener-b97fc8746-vlx9z\" (UID: \"d15b0450-e100-4ee7-ad62-88601eedc4f2\") " pod="openstack/barbican-keystone-listener-b97fc8746-vlx9z" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.610402 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d15b0450-e100-4ee7-ad62-88601eedc4f2-config-data\") pod \"barbican-keystone-listener-b97fc8746-vlx9z\" (UID: \"d15b0450-e100-4ee7-ad62-88601eedc4f2\") " pod="openstack/barbican-keystone-listener-b97fc8746-vlx9z" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.622680 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtknw\" (UniqueName: \"kubernetes.io/projected/d15b0450-e100-4ee7-ad62-88601eedc4f2-kube-api-access-vtknw\") pod \"barbican-keystone-listener-b97fc8746-vlx9z\" (UID: \"d15b0450-e100-4ee7-ad62-88601eedc4f2\") " pod="openstack/barbican-keystone-listener-b97fc8746-vlx9z" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.622939 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwzdf\" (UniqueName: \"kubernetes.io/projected/19a55666-374f-430a-a80b-f61538cbfb6b-kube-api-access-qwzdf\") pod \"dnsmasq-dns-867d6b4989-5mkmw\" (UID: \"19a55666-374f-430a-a80b-f61538cbfb6b\") " pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.623115 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6bf48f9d67-g95nd" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.655035 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-b97fc8746-vlx9z" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.702541 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8799f51-9cec-454b-bdbe-860ce455baca-combined-ca-bundle\") pod \"barbican-api-57c45ff45d-ghdvr\" (UID: \"a8799f51-9cec-454b-bdbe-860ce455baca\") " pod="openstack/barbican-api-57c45ff45d-ghdvr" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.702674 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckbgc\" (UniqueName: \"kubernetes.io/projected/a8799f51-9cec-454b-bdbe-860ce455baca-kube-api-access-ckbgc\") pod \"barbican-api-57c45ff45d-ghdvr\" (UID: \"a8799f51-9cec-454b-bdbe-860ce455baca\") " pod="openstack/barbican-api-57c45ff45d-ghdvr" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.702700 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8799f51-9cec-454b-bdbe-860ce455baca-logs\") pod \"barbican-api-57c45ff45d-ghdvr\" (UID: \"a8799f51-9cec-454b-bdbe-860ce455baca\") " pod="openstack/barbican-api-57c45ff45d-ghdvr" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.702728 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a8799f51-9cec-454b-bdbe-860ce455baca-config-data-custom\") pod \"barbican-api-57c45ff45d-ghdvr\" (UID: \"a8799f51-9cec-454b-bdbe-860ce455baca\") " pod="openstack/barbican-api-57c45ff45d-ghdvr" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.702750 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8799f51-9cec-454b-bdbe-860ce455baca-config-data\") pod \"barbican-api-57c45ff45d-ghdvr\" (UID: \"a8799f51-9cec-454b-bdbe-860ce455baca\") " pod="openstack/barbican-api-57c45ff45d-ghdvr" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.715155 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.804129 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckbgc\" (UniqueName: \"kubernetes.io/projected/a8799f51-9cec-454b-bdbe-860ce455baca-kube-api-access-ckbgc\") pod \"barbican-api-57c45ff45d-ghdvr\" (UID: \"a8799f51-9cec-454b-bdbe-860ce455baca\") " pod="openstack/barbican-api-57c45ff45d-ghdvr" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.804184 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8799f51-9cec-454b-bdbe-860ce455baca-logs\") pod \"barbican-api-57c45ff45d-ghdvr\" (UID: \"a8799f51-9cec-454b-bdbe-860ce455baca\") " pod="openstack/barbican-api-57c45ff45d-ghdvr" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.804220 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a8799f51-9cec-454b-bdbe-860ce455baca-config-data-custom\") pod \"barbican-api-57c45ff45d-ghdvr\" (UID: \"a8799f51-9cec-454b-bdbe-860ce455baca\") " pod="openstack/barbican-api-57c45ff45d-ghdvr" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.804250 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8799f51-9cec-454b-bdbe-860ce455baca-config-data\") pod \"barbican-api-57c45ff45d-ghdvr\" (UID: \"a8799f51-9cec-454b-bdbe-860ce455baca\") " pod="openstack/barbican-api-57c45ff45d-ghdvr" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.804309 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8799f51-9cec-454b-bdbe-860ce455baca-combined-ca-bundle\") pod \"barbican-api-57c45ff45d-ghdvr\" (UID: \"a8799f51-9cec-454b-bdbe-860ce455baca\") " pod="openstack/barbican-api-57c45ff45d-ghdvr" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.805413 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8799f51-9cec-454b-bdbe-860ce455baca-logs\") pod \"barbican-api-57c45ff45d-ghdvr\" (UID: \"a8799f51-9cec-454b-bdbe-860ce455baca\") " pod="openstack/barbican-api-57c45ff45d-ghdvr" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.811154 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8799f51-9cec-454b-bdbe-860ce455baca-combined-ca-bundle\") pod \"barbican-api-57c45ff45d-ghdvr\" (UID: \"a8799f51-9cec-454b-bdbe-860ce455baca\") " pod="openstack/barbican-api-57c45ff45d-ghdvr" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.823632 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a8799f51-9cec-454b-bdbe-860ce455baca-config-data-custom\") pod \"barbican-api-57c45ff45d-ghdvr\" (UID: \"a8799f51-9cec-454b-bdbe-860ce455baca\") " pod="openstack/barbican-api-57c45ff45d-ghdvr" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.824694 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8799f51-9cec-454b-bdbe-860ce455baca-config-data\") pod \"barbican-api-57c45ff45d-ghdvr\" (UID: \"a8799f51-9cec-454b-bdbe-860ce455baca\") " pod="openstack/barbican-api-57c45ff45d-ghdvr" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.826781 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckbgc\" (UniqueName: \"kubernetes.io/projected/a8799f51-9cec-454b-bdbe-860ce455baca-kube-api-access-ckbgc\") pod \"barbican-api-57c45ff45d-ghdvr\" (UID: \"a8799f51-9cec-454b-bdbe-860ce455baca\") " pod="openstack/barbican-api-57c45ff45d-ghdvr" Oct 02 10:07:46 crc kubenswrapper[4934]: I1002 10:07:46.902926 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57c45ff45d-ghdvr" Oct 02 10:07:47 crc kubenswrapper[4934]: I1002 10:07:47.137618 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6bf48f9d67-g95nd"] Oct 02 10:07:47 crc kubenswrapper[4934]: I1002 10:07:47.198498 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-b97fc8746-vlx9z"] Oct 02 10:07:47 crc kubenswrapper[4934]: I1002 10:07:47.287925 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-867d6b4989-5mkmw"] Oct 02 10:07:47 crc kubenswrapper[4934]: W1002 10:07:47.295669 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19a55666_374f_430a_a80b_f61538cbfb6b.slice/crio-ffc3eddf96decb3daad8467f1f87c325bcef77f1f66824faa9f0eab7ff4ddf18 WatchSource:0}: Error finding container ffc3eddf96decb3daad8467f1f87c325bcef77f1f66824faa9f0eab7ff4ddf18: Status 404 returned error can't find the container with id ffc3eddf96decb3daad8467f1f87c325bcef77f1f66824faa9f0eab7ff4ddf18 Oct 02 10:07:47 crc kubenswrapper[4934]: I1002 10:07:47.438919 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-57c45ff45d-ghdvr"] Oct 02 10:07:47 crc kubenswrapper[4934]: W1002 10:07:47.448440 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda8799f51_9cec_454b_bdbe_860ce455baca.slice/crio-425c56e98c096fa60ea560c5ecd4529df3356c6d41f8bf7052a06a014e538115 WatchSource:0}: Error finding container 425c56e98c096fa60ea560c5ecd4529df3356c6d41f8bf7052a06a014e538115: Status 404 returned error can't find the container with id 425c56e98c096fa60ea560c5ecd4529df3356c6d41f8bf7052a06a014e538115 Oct 02 10:07:48 crc kubenswrapper[4934]: I1002 10:07:48.082890 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6bf48f9d67-g95nd" event={"ID":"547c6c6d-0f01-4feb-acc6-2e3045407b64","Type":"ContainerStarted","Data":"fb60fab0d9fd014001fa386851aec18c4189763f270a448ffbeee76e27776329"} Oct 02 10:07:48 crc kubenswrapper[4934]: I1002 10:07:48.097753 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57c45ff45d-ghdvr" event={"ID":"a8799f51-9cec-454b-bdbe-860ce455baca","Type":"ContainerStarted","Data":"59ff67fae05185b501e4f0db278a33b889fc417ce780a16ef043c8e691f5fcbd"} Oct 02 10:07:48 crc kubenswrapper[4934]: I1002 10:07:48.097825 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57c45ff45d-ghdvr" event={"ID":"a8799f51-9cec-454b-bdbe-860ce455baca","Type":"ContainerStarted","Data":"ee4233b97847dce506924805dd7b618d62016dea5d61be04baec7613cd33b597"} Oct 02 10:07:48 crc kubenswrapper[4934]: I1002 10:07:48.097837 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57c45ff45d-ghdvr" event={"ID":"a8799f51-9cec-454b-bdbe-860ce455baca","Type":"ContainerStarted","Data":"425c56e98c096fa60ea560c5ecd4529df3356c6d41f8bf7052a06a014e538115"} Oct 02 10:07:48 crc kubenswrapper[4934]: I1002 10:07:48.098883 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-57c45ff45d-ghdvr" Oct 02 10:07:48 crc kubenswrapper[4934]: I1002 10:07:48.098904 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-57c45ff45d-ghdvr" Oct 02 10:07:48 crc kubenswrapper[4934]: I1002 10:07:48.105404 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-b97fc8746-vlx9z" event={"ID":"d15b0450-e100-4ee7-ad62-88601eedc4f2","Type":"ContainerStarted","Data":"7eab5abdc57629a153a839021a55ef264408229deaf5d32392f17de6def77d2f"} Oct 02 10:07:48 crc kubenswrapper[4934]: I1002 10:07:48.112055 4934 generic.go:334] "Generic (PLEG): container finished" podID="19a55666-374f-430a-a80b-f61538cbfb6b" containerID="a6d9a0ef8d05ad65a820fdc35f28c79f4ed00750f5d12b23a12ee5aa69aac5fe" exitCode=0 Oct 02 10:07:48 crc kubenswrapper[4934]: I1002 10:07:48.113724 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" event={"ID":"19a55666-374f-430a-a80b-f61538cbfb6b","Type":"ContainerDied","Data":"a6d9a0ef8d05ad65a820fdc35f28c79f4ed00750f5d12b23a12ee5aa69aac5fe"} Oct 02 10:07:48 crc kubenswrapper[4934]: I1002 10:07:48.113844 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" event={"ID":"19a55666-374f-430a-a80b-f61538cbfb6b","Type":"ContainerStarted","Data":"ffc3eddf96decb3daad8467f1f87c325bcef77f1f66824faa9f0eab7ff4ddf18"} Oct 02 10:07:48 crc kubenswrapper[4934]: I1002 10:07:48.123893 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-57c45ff45d-ghdvr" podStartSLOduration=2.123874048 podStartE2EDuration="2.123874048s" podCreationTimestamp="2025-10-02 10:07:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:07:48.120478531 +0000 UTC m=+1139.873120053" watchObservedRunningTime="2025-10-02 10:07:48.123874048 +0000 UTC m=+1139.876515570" Oct 02 10:07:49 crc kubenswrapper[4934]: I1002 10:07:49.131861 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" event={"ID":"19a55666-374f-430a-a80b-f61538cbfb6b","Type":"ContainerStarted","Data":"674a753da51398eb434068655d7cd432bb332bd68386e5babb0442791a8674ab"} Oct 02 10:07:49 crc kubenswrapper[4934]: I1002 10:07:49.134034 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" Oct 02 10:07:49 crc kubenswrapper[4934]: I1002 10:07:49.134776 4934 generic.go:334] "Generic (PLEG): container finished" podID="087ad031-cdf0-45e4-b1cf-ae777e8cc5fb" containerID="1cd0dc5312e4c3b419afaad3821f775ed69083d11a25939daa211e3938162c93" exitCode=0 Oct 02 10:07:49 crc kubenswrapper[4934]: I1002 10:07:49.135023 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-kb5lq" event={"ID":"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb","Type":"ContainerDied","Data":"1cd0dc5312e4c3b419afaad3821f775ed69083d11a25939daa211e3938162c93"} Oct 02 10:07:49 crc kubenswrapper[4934]: I1002 10:07:49.171243 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" podStartSLOduration=3.17122171 podStartE2EDuration="3.17122171s" podCreationTimestamp="2025-10-02 10:07:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:07:49.148839772 +0000 UTC m=+1140.901481294" watchObservedRunningTime="2025-10-02 10:07:49.17122171 +0000 UTC m=+1140.923863232" Oct 02 10:07:49 crc kubenswrapper[4934]: I1002 10:07:49.255284 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:07:49 crc kubenswrapper[4934]: I1002 10:07:49.801152 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-8648b56f4-5f8xk"] Oct 02 10:07:49 crc kubenswrapper[4934]: I1002 10:07:49.803241 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:07:49 crc kubenswrapper[4934]: I1002 10:07:49.810987 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 02 10:07:49 crc kubenswrapper[4934]: I1002 10:07:49.811261 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 02 10:07:49 crc kubenswrapper[4934]: I1002 10:07:49.820492 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-8648b56f4-5f8xk"] Oct 02 10:07:49 crc kubenswrapper[4934]: I1002 10:07:49.985401 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-config-data\") pod \"barbican-api-8648b56f4-5f8xk\" (UID: \"350928d4-5a0b-4331-b1c2-220517d313a5\") " pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:07:49 crc kubenswrapper[4934]: I1002 10:07:49.985745 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-combined-ca-bundle\") pod \"barbican-api-8648b56f4-5f8xk\" (UID: \"350928d4-5a0b-4331-b1c2-220517d313a5\") " pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:07:49 crc kubenswrapper[4934]: I1002 10:07:49.985946 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/350928d4-5a0b-4331-b1c2-220517d313a5-logs\") pod \"barbican-api-8648b56f4-5f8xk\" (UID: \"350928d4-5a0b-4331-b1c2-220517d313a5\") " pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:07:49 crc kubenswrapper[4934]: I1002 10:07:49.986094 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-public-tls-certs\") pod \"barbican-api-8648b56f4-5f8xk\" (UID: \"350928d4-5a0b-4331-b1c2-220517d313a5\") " pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:07:49 crc kubenswrapper[4934]: I1002 10:07:49.986275 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-config-data-custom\") pod \"barbican-api-8648b56f4-5f8xk\" (UID: \"350928d4-5a0b-4331-b1c2-220517d313a5\") " pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:07:49 crc kubenswrapper[4934]: I1002 10:07:49.986441 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmn6q\" (UniqueName: \"kubernetes.io/projected/350928d4-5a0b-4331-b1c2-220517d313a5-kube-api-access-mmn6q\") pod \"barbican-api-8648b56f4-5f8xk\" (UID: \"350928d4-5a0b-4331-b1c2-220517d313a5\") " pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:07:49 crc kubenswrapper[4934]: I1002 10:07:49.986734 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-internal-tls-certs\") pod \"barbican-api-8648b56f4-5f8xk\" (UID: \"350928d4-5a0b-4331-b1c2-220517d313a5\") " pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.088031 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-config-data\") pod \"barbican-api-8648b56f4-5f8xk\" (UID: \"350928d4-5a0b-4331-b1c2-220517d313a5\") " pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.088220 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-combined-ca-bundle\") pod \"barbican-api-8648b56f4-5f8xk\" (UID: \"350928d4-5a0b-4331-b1c2-220517d313a5\") " pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.088332 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/350928d4-5a0b-4331-b1c2-220517d313a5-logs\") pod \"barbican-api-8648b56f4-5f8xk\" (UID: \"350928d4-5a0b-4331-b1c2-220517d313a5\") " pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.088427 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-public-tls-certs\") pod \"barbican-api-8648b56f4-5f8xk\" (UID: \"350928d4-5a0b-4331-b1c2-220517d313a5\") " pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.088546 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-config-data-custom\") pod \"barbican-api-8648b56f4-5f8xk\" (UID: \"350928d4-5a0b-4331-b1c2-220517d313a5\") " pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.088674 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmn6q\" (UniqueName: \"kubernetes.io/projected/350928d4-5a0b-4331-b1c2-220517d313a5-kube-api-access-mmn6q\") pod \"barbican-api-8648b56f4-5f8xk\" (UID: \"350928d4-5a0b-4331-b1c2-220517d313a5\") " pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.088834 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-internal-tls-certs\") pod \"barbican-api-8648b56f4-5f8xk\" (UID: \"350928d4-5a0b-4331-b1c2-220517d313a5\") " pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.089335 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/350928d4-5a0b-4331-b1c2-220517d313a5-logs\") pod \"barbican-api-8648b56f4-5f8xk\" (UID: \"350928d4-5a0b-4331-b1c2-220517d313a5\") " pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.093138 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-internal-tls-certs\") pod \"barbican-api-8648b56f4-5f8xk\" (UID: \"350928d4-5a0b-4331-b1c2-220517d313a5\") " pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.108667 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-config-data-custom\") pod \"barbican-api-8648b56f4-5f8xk\" (UID: \"350928d4-5a0b-4331-b1c2-220517d313a5\") " pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.109629 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-public-tls-certs\") pod \"barbican-api-8648b56f4-5f8xk\" (UID: \"350928d4-5a0b-4331-b1c2-220517d313a5\") " pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.110309 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-config-data\") pod \"barbican-api-8648b56f4-5f8xk\" (UID: \"350928d4-5a0b-4331-b1c2-220517d313a5\") " pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.111462 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-combined-ca-bundle\") pod \"barbican-api-8648b56f4-5f8xk\" (UID: \"350928d4-5a0b-4331-b1c2-220517d313a5\") " pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.111928 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmn6q\" (UniqueName: \"kubernetes.io/projected/350928d4-5a0b-4331-b1c2-220517d313a5-kube-api-access-mmn6q\") pod \"barbican-api-8648b56f4-5f8xk\" (UID: \"350928d4-5a0b-4331-b1c2-220517d313a5\") " pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.145800 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6bf48f9d67-g95nd" event={"ID":"547c6c6d-0f01-4feb-acc6-2e3045407b64","Type":"ContainerStarted","Data":"ca5c961136b99e80a8ae0bb9c203d870c7355071cc4a476bdb7c41606bdef8b7"} Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.157781 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-b97fc8746-vlx9z" event={"ID":"d15b0450-e100-4ee7-ad62-88601eedc4f2","Type":"ContainerStarted","Data":"c83811ac76f059e3b264882bb3a6859bb99aa9ea06694e42d9967733e316ac15"} Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.262837 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.447062 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-kb5lq" Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.597565 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-combined-ca-bundle\") pod \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\" (UID: \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\") " Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.597896 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-db-sync-config-data\") pod \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\" (UID: \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\") " Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.598008 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlskg\" (UniqueName: \"kubernetes.io/projected/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-kube-api-access-qlskg\") pod \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\" (UID: \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\") " Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.598054 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-scripts\") pod \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\" (UID: \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\") " Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.598083 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-etc-machine-id\") pod \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\" (UID: \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\") " Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.598120 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-config-data\") pod \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\" (UID: \"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb\") " Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.601687 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "087ad031-cdf0-45e4-b1cf-ae777e8cc5fb" (UID: "087ad031-cdf0-45e4-b1cf-ae777e8cc5fb"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.605080 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-kube-api-access-qlskg" (OuterVolumeSpecName: "kube-api-access-qlskg") pod "087ad031-cdf0-45e4-b1cf-ae777e8cc5fb" (UID: "087ad031-cdf0-45e4-b1cf-ae777e8cc5fb"). InnerVolumeSpecName "kube-api-access-qlskg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.605326 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "087ad031-cdf0-45e4-b1cf-ae777e8cc5fb" (UID: "087ad031-cdf0-45e4-b1cf-ae777e8cc5fb"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.613002 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-scripts" (OuterVolumeSpecName: "scripts") pod "087ad031-cdf0-45e4-b1cf-ae777e8cc5fb" (UID: "087ad031-cdf0-45e4-b1cf-ae777e8cc5fb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.648956 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "087ad031-cdf0-45e4-b1cf-ae777e8cc5fb" (UID: "087ad031-cdf0-45e4-b1cf-ae777e8cc5fb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.658798 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-config-data" (OuterVolumeSpecName: "config-data") pod "087ad031-cdf0-45e4-b1cf-ae777e8cc5fb" (UID: "087ad031-cdf0-45e4-b1cf-ae777e8cc5fb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.700130 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.700440 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.700530 4934 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.700623 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlskg\" (UniqueName: \"kubernetes.io/projected/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-kube-api-access-qlskg\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.700729 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.700836 4934 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:50 crc kubenswrapper[4934]: W1002 10:07:50.768164 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod350928d4_5a0b_4331_b1c2_220517d313a5.slice/crio-af20426c4cf3268848d21e7e02b8f5c6c16fdc7187bdb42848932be47df42728 WatchSource:0}: Error finding container af20426c4cf3268848d21e7e02b8f5c6c16fdc7187bdb42848932be47df42728: Status 404 returned error can't find the container with id af20426c4cf3268848d21e7e02b8f5c6c16fdc7187bdb42848932be47df42728 Oct 02 10:07:50 crc kubenswrapper[4934]: I1002 10:07:50.769916 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-8648b56f4-5f8xk"] Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.169852 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8648b56f4-5f8xk" event={"ID":"350928d4-5a0b-4331-b1c2-220517d313a5","Type":"ContainerStarted","Data":"9245f97a19f447f9859a60a1c0d5f3f95ccff037fb5235ac35a03f447f11628f"} Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.170334 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8648b56f4-5f8xk" event={"ID":"350928d4-5a0b-4331-b1c2-220517d313a5","Type":"ContainerStarted","Data":"af20426c4cf3268848d21e7e02b8f5c6c16fdc7187bdb42848932be47df42728"} Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.172227 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-kb5lq" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.173233 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-kb5lq" event={"ID":"087ad031-cdf0-45e4-b1cf-ae777e8cc5fb","Type":"ContainerDied","Data":"807d9943655205ef342496d4ab704cd5e64ab362985bac33828e63b29adb19f3"} Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.173273 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="807d9943655205ef342496d4ab704cd5e64ab362985bac33828e63b29adb19f3" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.180218 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6bf48f9d67-g95nd" event={"ID":"547c6c6d-0f01-4feb-acc6-2e3045407b64","Type":"ContainerStarted","Data":"c52d38070f0e71890f03b32d417b88c6062d9efcaddea407689d3ce9f0ed9989"} Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.183954 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-b97fc8746-vlx9z" event={"ID":"d15b0450-e100-4ee7-ad62-88601eedc4f2","Type":"ContainerStarted","Data":"2e7e1ea7602c5dbc0c4396593163b8efff5196b8cba864f827def060e37814e7"} Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.219298 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-6bf48f9d67-g95nd" podStartSLOduration=2.685917177 podStartE2EDuration="5.219279086s" podCreationTimestamp="2025-10-02 10:07:46 +0000 UTC" firstStartedPulling="2025-10-02 10:07:47.13986193 +0000 UTC m=+1138.892503442" lastFinishedPulling="2025-10-02 10:07:49.673223829 +0000 UTC m=+1141.425865351" observedRunningTime="2025-10-02 10:07:51.19940841 +0000 UTC m=+1142.952049962" watchObservedRunningTime="2025-10-02 10:07:51.219279086 +0000 UTC m=+1142.971920608" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.233298 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-b97fc8746-vlx9z" podStartSLOduration=2.767741978 podStartE2EDuration="5.233279455s" podCreationTimestamp="2025-10-02 10:07:46 +0000 UTC" firstStartedPulling="2025-10-02 10:07:47.209671939 +0000 UTC m=+1138.962313461" lastFinishedPulling="2025-10-02 10:07:49.675209416 +0000 UTC m=+1141.427850938" observedRunningTime="2025-10-02 10:07:51.221021565 +0000 UTC m=+1142.973663087" watchObservedRunningTime="2025-10-02 10:07:51.233279455 +0000 UTC m=+1142.985920977" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.454352 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 10:07:51 crc kubenswrapper[4934]: E1002 10:07:51.455043 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="087ad031-cdf0-45e4-b1cf-ae777e8cc5fb" containerName="cinder-db-sync" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.455449 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="087ad031-cdf0-45e4-b1cf-ae777e8cc5fb" containerName="cinder-db-sync" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.456152 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="087ad031-cdf0-45e4-b1cf-ae777e8cc5fb" containerName="cinder-db-sync" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.464623 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.470183 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.470361 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.470444 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-b6f84" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.470526 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.477616 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.528707 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-867d6b4989-5mkmw"] Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.532949 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" podUID="19a55666-374f-430a-a80b-f61538cbfb6b" containerName="dnsmasq-dns" containerID="cri-o://674a753da51398eb434068655d7cd432bb332bd68386e5babb0442791a8674ab" gracePeriod=10 Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.556230 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79fb47bfff-98d57"] Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.557719 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79fb47bfff-98d57" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.575043 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79fb47bfff-98d57"] Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.621571 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0d0eed2-217a-4dbc-906a-797c09b32081-scripts\") pod \"cinder-scheduler-0\" (UID: \"e0d0eed2-217a-4dbc-906a-797c09b32081\") " pod="openstack/cinder-scheduler-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.621633 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0d0eed2-217a-4dbc-906a-797c09b32081-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e0d0eed2-217a-4dbc-906a-797c09b32081\") " pod="openstack/cinder-scheduler-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.621709 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0d0eed2-217a-4dbc-906a-797c09b32081-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e0d0eed2-217a-4dbc-906a-797c09b32081\") " pod="openstack/cinder-scheduler-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.621725 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mt2t\" (UniqueName: \"kubernetes.io/projected/e0d0eed2-217a-4dbc-906a-797c09b32081-kube-api-access-8mt2t\") pod \"cinder-scheduler-0\" (UID: \"e0d0eed2-217a-4dbc-906a-797c09b32081\") " pod="openstack/cinder-scheduler-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.621789 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0d0eed2-217a-4dbc-906a-797c09b32081-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e0d0eed2-217a-4dbc-906a-797c09b32081\") " pod="openstack/cinder-scheduler-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.621829 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0d0eed2-217a-4dbc-906a-797c09b32081-config-data\") pod \"cinder-scheduler-0\" (UID: \"e0d0eed2-217a-4dbc-906a-797c09b32081\") " pod="openstack/cinder-scheduler-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.671649 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.673817 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.680404 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.694524 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.725371 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-ovsdbserver-nb\") pod \"dnsmasq-dns-79fb47bfff-98d57\" (UID: \"f55848b6-dd77-46d8-8366-4715042cdd5d\") " pod="openstack/dnsmasq-dns-79fb47bfff-98d57" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.725433 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0d0eed2-217a-4dbc-906a-797c09b32081-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e0d0eed2-217a-4dbc-906a-797c09b32081\") " pod="openstack/cinder-scheduler-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.725465 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0d0eed2-217a-4dbc-906a-797c09b32081-config-data\") pod \"cinder-scheduler-0\" (UID: \"e0d0eed2-217a-4dbc-906a-797c09b32081\") " pod="openstack/cinder-scheduler-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.725492 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0d0eed2-217a-4dbc-906a-797c09b32081-scripts\") pod \"cinder-scheduler-0\" (UID: \"e0d0eed2-217a-4dbc-906a-797c09b32081\") " pod="openstack/cinder-scheduler-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.725520 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0d0eed2-217a-4dbc-906a-797c09b32081-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e0d0eed2-217a-4dbc-906a-797c09b32081\") " pod="openstack/cinder-scheduler-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.725542 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-dns-svc\") pod \"dnsmasq-dns-79fb47bfff-98d57\" (UID: \"f55848b6-dd77-46d8-8366-4715042cdd5d\") " pod="openstack/dnsmasq-dns-79fb47bfff-98d57" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.725587 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-config\") pod \"dnsmasq-dns-79fb47bfff-98d57\" (UID: \"f55848b6-dd77-46d8-8366-4715042cdd5d\") " pod="openstack/dnsmasq-dns-79fb47bfff-98d57" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.725622 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-ovsdbserver-sb\") pod \"dnsmasq-dns-79fb47bfff-98d57\" (UID: \"f55848b6-dd77-46d8-8366-4715042cdd5d\") " pod="openstack/dnsmasq-dns-79fb47bfff-98d57" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.725907 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-dns-swift-storage-0\") pod \"dnsmasq-dns-79fb47bfff-98d57\" (UID: \"f55848b6-dd77-46d8-8366-4715042cdd5d\") " pod="openstack/dnsmasq-dns-79fb47bfff-98d57" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.725925 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc2vp\" (UniqueName: \"kubernetes.io/projected/f55848b6-dd77-46d8-8366-4715042cdd5d-kube-api-access-dc2vp\") pod \"dnsmasq-dns-79fb47bfff-98d57\" (UID: \"f55848b6-dd77-46d8-8366-4715042cdd5d\") " pod="openstack/dnsmasq-dns-79fb47bfff-98d57" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.725946 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0d0eed2-217a-4dbc-906a-797c09b32081-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e0d0eed2-217a-4dbc-906a-797c09b32081\") " pod="openstack/cinder-scheduler-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.725963 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mt2t\" (UniqueName: \"kubernetes.io/projected/e0d0eed2-217a-4dbc-906a-797c09b32081-kube-api-access-8mt2t\") pod \"cinder-scheduler-0\" (UID: \"e0d0eed2-217a-4dbc-906a-797c09b32081\") " pod="openstack/cinder-scheduler-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.727927 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0d0eed2-217a-4dbc-906a-797c09b32081-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e0d0eed2-217a-4dbc-906a-797c09b32081\") " pod="openstack/cinder-scheduler-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.738288 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0d0eed2-217a-4dbc-906a-797c09b32081-scripts\") pod \"cinder-scheduler-0\" (UID: \"e0d0eed2-217a-4dbc-906a-797c09b32081\") " pod="openstack/cinder-scheduler-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.745147 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0d0eed2-217a-4dbc-906a-797c09b32081-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e0d0eed2-217a-4dbc-906a-797c09b32081\") " pod="openstack/cinder-scheduler-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.746040 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0d0eed2-217a-4dbc-906a-797c09b32081-config-data\") pod \"cinder-scheduler-0\" (UID: \"e0d0eed2-217a-4dbc-906a-797c09b32081\") " pod="openstack/cinder-scheduler-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.749518 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0d0eed2-217a-4dbc-906a-797c09b32081-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e0d0eed2-217a-4dbc-906a-797c09b32081\") " pod="openstack/cinder-scheduler-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.758716 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mt2t\" (UniqueName: \"kubernetes.io/projected/e0d0eed2-217a-4dbc-906a-797c09b32081-kube-api-access-8mt2t\") pod \"cinder-scheduler-0\" (UID: \"e0d0eed2-217a-4dbc-906a-797c09b32081\") " pod="openstack/cinder-scheduler-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.787818 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.833486 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6807b23-4962-4a52-8905-af632dbf9b5c-config-data\") pod \"cinder-api-0\" (UID: \"a6807b23-4962-4a52-8905-af632dbf9b5c\") " pod="openstack/cinder-api-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.833546 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a6807b23-4962-4a52-8905-af632dbf9b5c-config-data-custom\") pod \"cinder-api-0\" (UID: \"a6807b23-4962-4a52-8905-af632dbf9b5c\") " pod="openstack/cinder-api-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.833586 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-dns-svc\") pod \"dnsmasq-dns-79fb47bfff-98d57\" (UID: \"f55848b6-dd77-46d8-8366-4715042cdd5d\") " pod="openstack/dnsmasq-dns-79fb47bfff-98d57" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.833610 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6807b23-4962-4a52-8905-af632dbf9b5c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a6807b23-4962-4a52-8905-af632dbf9b5c\") " pod="openstack/cinder-api-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.833674 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-config\") pod \"dnsmasq-dns-79fb47bfff-98d57\" (UID: \"f55848b6-dd77-46d8-8366-4715042cdd5d\") " pod="openstack/dnsmasq-dns-79fb47bfff-98d57" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.833697 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6807b23-4962-4a52-8905-af632dbf9b5c-scripts\") pod \"cinder-api-0\" (UID: \"a6807b23-4962-4a52-8905-af632dbf9b5c\") " pod="openstack/cinder-api-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.833734 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-ovsdbserver-sb\") pod \"dnsmasq-dns-79fb47bfff-98d57\" (UID: \"f55848b6-dd77-46d8-8366-4715042cdd5d\") " pod="openstack/dnsmasq-dns-79fb47bfff-98d57" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.833754 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-dns-swift-storage-0\") pod \"dnsmasq-dns-79fb47bfff-98d57\" (UID: \"f55848b6-dd77-46d8-8366-4715042cdd5d\") " pod="openstack/dnsmasq-dns-79fb47bfff-98d57" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.833772 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc2vp\" (UniqueName: \"kubernetes.io/projected/f55848b6-dd77-46d8-8366-4715042cdd5d-kube-api-access-dc2vp\") pod \"dnsmasq-dns-79fb47bfff-98d57\" (UID: \"f55848b6-dd77-46d8-8366-4715042cdd5d\") " pod="openstack/dnsmasq-dns-79fb47bfff-98d57" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.833808 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdvdz\" (UniqueName: \"kubernetes.io/projected/a6807b23-4962-4a52-8905-af632dbf9b5c-kube-api-access-mdvdz\") pod \"cinder-api-0\" (UID: \"a6807b23-4962-4a52-8905-af632dbf9b5c\") " pod="openstack/cinder-api-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.833847 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-ovsdbserver-nb\") pod \"dnsmasq-dns-79fb47bfff-98d57\" (UID: \"f55848b6-dd77-46d8-8366-4715042cdd5d\") " pod="openstack/dnsmasq-dns-79fb47bfff-98d57" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.833875 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a6807b23-4962-4a52-8905-af632dbf9b5c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a6807b23-4962-4a52-8905-af632dbf9b5c\") " pod="openstack/cinder-api-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.833898 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6807b23-4962-4a52-8905-af632dbf9b5c-logs\") pod \"cinder-api-0\" (UID: \"a6807b23-4962-4a52-8905-af632dbf9b5c\") " pod="openstack/cinder-api-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.834794 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-dns-svc\") pod \"dnsmasq-dns-79fb47bfff-98d57\" (UID: \"f55848b6-dd77-46d8-8366-4715042cdd5d\") " pod="openstack/dnsmasq-dns-79fb47bfff-98d57" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.835337 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-config\") pod \"dnsmasq-dns-79fb47bfff-98d57\" (UID: \"f55848b6-dd77-46d8-8366-4715042cdd5d\") " pod="openstack/dnsmasq-dns-79fb47bfff-98d57" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.835911 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-ovsdbserver-sb\") pod \"dnsmasq-dns-79fb47bfff-98d57\" (UID: \"f55848b6-dd77-46d8-8366-4715042cdd5d\") " pod="openstack/dnsmasq-dns-79fb47bfff-98d57" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.836410 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-dns-swift-storage-0\") pod \"dnsmasq-dns-79fb47bfff-98d57\" (UID: \"f55848b6-dd77-46d8-8366-4715042cdd5d\") " pod="openstack/dnsmasq-dns-79fb47bfff-98d57" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.837111 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-ovsdbserver-nb\") pod \"dnsmasq-dns-79fb47bfff-98d57\" (UID: \"f55848b6-dd77-46d8-8366-4715042cdd5d\") " pod="openstack/dnsmasq-dns-79fb47bfff-98d57" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.872490 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc2vp\" (UniqueName: \"kubernetes.io/projected/f55848b6-dd77-46d8-8366-4715042cdd5d-kube-api-access-dc2vp\") pod \"dnsmasq-dns-79fb47bfff-98d57\" (UID: \"f55848b6-dd77-46d8-8366-4715042cdd5d\") " pod="openstack/dnsmasq-dns-79fb47bfff-98d57" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.921471 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79fb47bfff-98d57" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.936041 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdvdz\" (UniqueName: \"kubernetes.io/projected/a6807b23-4962-4a52-8905-af632dbf9b5c-kube-api-access-mdvdz\") pod \"cinder-api-0\" (UID: \"a6807b23-4962-4a52-8905-af632dbf9b5c\") " pod="openstack/cinder-api-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.936140 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a6807b23-4962-4a52-8905-af632dbf9b5c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a6807b23-4962-4a52-8905-af632dbf9b5c\") " pod="openstack/cinder-api-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.936494 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6807b23-4962-4a52-8905-af632dbf9b5c-logs\") pod \"cinder-api-0\" (UID: \"a6807b23-4962-4a52-8905-af632dbf9b5c\") " pod="openstack/cinder-api-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.936558 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a6807b23-4962-4a52-8905-af632dbf9b5c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a6807b23-4962-4a52-8905-af632dbf9b5c\") " pod="openstack/cinder-api-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.937070 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6807b23-4962-4a52-8905-af632dbf9b5c-logs\") pod \"cinder-api-0\" (UID: \"a6807b23-4962-4a52-8905-af632dbf9b5c\") " pod="openstack/cinder-api-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.937124 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6807b23-4962-4a52-8905-af632dbf9b5c-config-data\") pod \"cinder-api-0\" (UID: \"a6807b23-4962-4a52-8905-af632dbf9b5c\") " pod="openstack/cinder-api-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.937175 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a6807b23-4962-4a52-8905-af632dbf9b5c-config-data-custom\") pod \"cinder-api-0\" (UID: \"a6807b23-4962-4a52-8905-af632dbf9b5c\") " pod="openstack/cinder-api-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.937203 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6807b23-4962-4a52-8905-af632dbf9b5c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a6807b23-4962-4a52-8905-af632dbf9b5c\") " pod="openstack/cinder-api-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.937246 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6807b23-4962-4a52-8905-af632dbf9b5c-scripts\") pod \"cinder-api-0\" (UID: \"a6807b23-4962-4a52-8905-af632dbf9b5c\") " pod="openstack/cinder-api-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.944285 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6807b23-4962-4a52-8905-af632dbf9b5c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a6807b23-4962-4a52-8905-af632dbf9b5c\") " pod="openstack/cinder-api-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.959214 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdvdz\" (UniqueName: \"kubernetes.io/projected/a6807b23-4962-4a52-8905-af632dbf9b5c-kube-api-access-mdvdz\") pod \"cinder-api-0\" (UID: \"a6807b23-4962-4a52-8905-af632dbf9b5c\") " pod="openstack/cinder-api-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.960511 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6807b23-4962-4a52-8905-af632dbf9b5c-scripts\") pod \"cinder-api-0\" (UID: \"a6807b23-4962-4a52-8905-af632dbf9b5c\") " pod="openstack/cinder-api-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.962659 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6807b23-4962-4a52-8905-af632dbf9b5c-config-data\") pod \"cinder-api-0\" (UID: \"a6807b23-4962-4a52-8905-af632dbf9b5c\") " pod="openstack/cinder-api-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.975730 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a6807b23-4962-4a52-8905-af632dbf9b5c-config-data-custom\") pod \"cinder-api-0\" (UID: \"a6807b23-4962-4a52-8905-af632dbf9b5c\") " pod="openstack/cinder-api-0" Oct 02 10:07:51 crc kubenswrapper[4934]: I1002 10:07:51.996162 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.089560 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.200292 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8648b56f4-5f8xk" event={"ID":"350928d4-5a0b-4331-b1c2-220517d313a5","Type":"ContainerStarted","Data":"41b55b90784f0d27b65abc25b560172adc6a97c8f0800218d04b4ab6f18fc610"} Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.200537 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.200621 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.220057 4934 generic.go:334] "Generic (PLEG): container finished" podID="19a55666-374f-430a-a80b-f61538cbfb6b" containerID="674a753da51398eb434068655d7cd432bb332bd68386e5babb0442791a8674ab" exitCode=0 Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.220419 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.220735 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" event={"ID":"19a55666-374f-430a-a80b-f61538cbfb6b","Type":"ContainerDied","Data":"674a753da51398eb434068655d7cd432bb332bd68386e5babb0442791a8674ab"} Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.220763 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867d6b4989-5mkmw" event={"ID":"19a55666-374f-430a-a80b-f61538cbfb6b","Type":"ContainerDied","Data":"ffc3eddf96decb3daad8467f1f87c325bcef77f1f66824faa9f0eab7ff4ddf18"} Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.220780 4934 scope.go:117] "RemoveContainer" containerID="674a753da51398eb434068655d7cd432bb332bd68386e5babb0442791a8674ab" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.234493 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-8648b56f4-5f8xk" podStartSLOduration=3.234474792 podStartE2EDuration="3.234474792s" podCreationTimestamp="2025-10-02 10:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:07:52.23197336 +0000 UTC m=+1143.984614892" watchObservedRunningTime="2025-10-02 10:07:52.234474792 +0000 UTC m=+1143.987116314" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.243055 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-ovsdbserver-nb\") pod \"19a55666-374f-430a-a80b-f61538cbfb6b\" (UID: \"19a55666-374f-430a-a80b-f61538cbfb6b\") " Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.243418 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-dns-swift-storage-0\") pod \"19a55666-374f-430a-a80b-f61538cbfb6b\" (UID: \"19a55666-374f-430a-a80b-f61538cbfb6b\") " Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.243443 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-ovsdbserver-sb\") pod \"19a55666-374f-430a-a80b-f61538cbfb6b\" (UID: \"19a55666-374f-430a-a80b-f61538cbfb6b\") " Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.243483 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwzdf\" (UniqueName: \"kubernetes.io/projected/19a55666-374f-430a-a80b-f61538cbfb6b-kube-api-access-qwzdf\") pod \"19a55666-374f-430a-a80b-f61538cbfb6b\" (UID: \"19a55666-374f-430a-a80b-f61538cbfb6b\") " Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.243508 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-config\") pod \"19a55666-374f-430a-a80b-f61538cbfb6b\" (UID: \"19a55666-374f-430a-a80b-f61538cbfb6b\") " Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.243538 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-dns-svc\") pod \"19a55666-374f-430a-a80b-f61538cbfb6b\" (UID: \"19a55666-374f-430a-a80b-f61538cbfb6b\") " Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.259018 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19a55666-374f-430a-a80b-f61538cbfb6b-kube-api-access-qwzdf" (OuterVolumeSpecName: "kube-api-access-qwzdf") pod "19a55666-374f-430a-a80b-f61538cbfb6b" (UID: "19a55666-374f-430a-a80b-f61538cbfb6b"). InnerVolumeSpecName "kube-api-access-qwzdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.275564 4934 scope.go:117] "RemoveContainer" containerID="a6d9a0ef8d05ad65a820fdc35f28c79f4ed00750f5d12b23a12ee5aa69aac5fe" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.315015 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-config" (OuterVolumeSpecName: "config") pod "19a55666-374f-430a-a80b-f61538cbfb6b" (UID: "19a55666-374f-430a-a80b-f61538cbfb6b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.325216 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "19a55666-374f-430a-a80b-f61538cbfb6b" (UID: "19a55666-374f-430a-a80b-f61538cbfb6b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.327924 4934 scope.go:117] "RemoveContainer" containerID="674a753da51398eb434068655d7cd432bb332bd68386e5babb0442791a8674ab" Oct 02 10:07:52 crc kubenswrapper[4934]: E1002 10:07:52.331679 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"674a753da51398eb434068655d7cd432bb332bd68386e5babb0442791a8674ab\": container with ID starting with 674a753da51398eb434068655d7cd432bb332bd68386e5babb0442791a8674ab not found: ID does not exist" containerID="674a753da51398eb434068655d7cd432bb332bd68386e5babb0442791a8674ab" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.331718 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"674a753da51398eb434068655d7cd432bb332bd68386e5babb0442791a8674ab"} err="failed to get container status \"674a753da51398eb434068655d7cd432bb332bd68386e5babb0442791a8674ab\": rpc error: code = NotFound desc = could not find container \"674a753da51398eb434068655d7cd432bb332bd68386e5babb0442791a8674ab\": container with ID starting with 674a753da51398eb434068655d7cd432bb332bd68386e5babb0442791a8674ab not found: ID does not exist" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.331743 4934 scope.go:117] "RemoveContainer" containerID="a6d9a0ef8d05ad65a820fdc35f28c79f4ed00750f5d12b23a12ee5aa69aac5fe" Oct 02 10:07:52 crc kubenswrapper[4934]: E1002 10:07:52.334824 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6d9a0ef8d05ad65a820fdc35f28c79f4ed00750f5d12b23a12ee5aa69aac5fe\": container with ID starting with a6d9a0ef8d05ad65a820fdc35f28c79f4ed00750f5d12b23a12ee5aa69aac5fe not found: ID does not exist" containerID="a6d9a0ef8d05ad65a820fdc35f28c79f4ed00750f5d12b23a12ee5aa69aac5fe" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.334873 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6d9a0ef8d05ad65a820fdc35f28c79f4ed00750f5d12b23a12ee5aa69aac5fe"} err="failed to get container status \"a6d9a0ef8d05ad65a820fdc35f28c79f4ed00750f5d12b23a12ee5aa69aac5fe\": rpc error: code = NotFound desc = could not find container \"a6d9a0ef8d05ad65a820fdc35f28c79f4ed00750f5d12b23a12ee5aa69aac5fe\": container with ID starting with a6d9a0ef8d05ad65a820fdc35f28c79f4ed00750f5d12b23a12ee5aa69aac5fe not found: ID does not exist" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.342049 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.347153 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.347172 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwzdf\" (UniqueName: \"kubernetes.io/projected/19a55666-374f-430a-a80b-f61538cbfb6b-kube-api-access-qwzdf\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.347181 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.358996 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "19a55666-374f-430a-a80b-f61538cbfb6b" (UID: "19a55666-374f-430a-a80b-f61538cbfb6b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.371394 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "19a55666-374f-430a-a80b-f61538cbfb6b" (UID: "19a55666-374f-430a-a80b-f61538cbfb6b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.381148 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "19a55666-374f-430a-a80b-f61538cbfb6b" (UID: "19a55666-374f-430a-a80b-f61538cbfb6b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:07:52 crc kubenswrapper[4934]: W1002 10:07:52.452837 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf55848b6_dd77_46d8_8366_4715042cdd5d.slice/crio-18645ce4ec2ad743abcf2b1feee55c89fe461a48841967d4b1116a5aefbce090 WatchSource:0}: Error finding container 18645ce4ec2ad743abcf2b1feee55c89fe461a48841967d4b1116a5aefbce090: Status 404 returned error can't find the container with id 18645ce4ec2ad743abcf2b1feee55c89fe461a48841967d4b1116a5aefbce090 Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.455951 4934 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.455994 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.456006 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19a55666-374f-430a-a80b-f61538cbfb6b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.461064 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79fb47bfff-98d57"] Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.557838 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 10:07:52 crc kubenswrapper[4934]: W1002 10:07:52.583048 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6807b23_4962_4a52_8905_af632dbf9b5c.slice/crio-5ab2b91762ee0df2e77c6a1f5fabf370779d0fac21eb66b99279e1a4ad76a03d WatchSource:0}: Error finding container 5ab2b91762ee0df2e77c6a1f5fabf370779d0fac21eb66b99279e1a4ad76a03d: Status 404 returned error can't find the container with id 5ab2b91762ee0df2e77c6a1f5fabf370779d0fac21eb66b99279e1a4ad76a03d Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.588914 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-867d6b4989-5mkmw"] Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.615760 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-867d6b4989-5mkmw"] Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.737863 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 02 10:07:52 crc kubenswrapper[4934]: E1002 10:07:52.738409 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19a55666-374f-430a-a80b-f61538cbfb6b" containerName="init" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.738431 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="19a55666-374f-430a-a80b-f61538cbfb6b" containerName="init" Oct 02 10:07:52 crc kubenswrapper[4934]: E1002 10:07:52.738470 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19a55666-374f-430a-a80b-f61538cbfb6b" containerName="dnsmasq-dns" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.738479 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="19a55666-374f-430a-a80b-f61538cbfb6b" containerName="dnsmasq-dns" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.738754 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="19a55666-374f-430a-a80b-f61538cbfb6b" containerName="dnsmasq-dns" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.740200 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.744045 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.744295 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-rbw29" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.744641 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.752011 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.866117 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/eefa17c1-494a-446b-8c20-96dc0886b88f-openstack-config-secret\") pod \"openstackclient\" (UID: \"eefa17c1-494a-446b-8c20-96dc0886b88f\") " pod="openstack/openstackclient" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.866199 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfg7l\" (UniqueName: \"kubernetes.io/projected/eefa17c1-494a-446b-8c20-96dc0886b88f-kube-api-access-gfg7l\") pod \"openstackclient\" (UID: \"eefa17c1-494a-446b-8c20-96dc0886b88f\") " pod="openstack/openstackclient" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.866278 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eefa17c1-494a-446b-8c20-96dc0886b88f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"eefa17c1-494a-446b-8c20-96dc0886b88f\") " pod="openstack/openstackclient" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.866570 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/eefa17c1-494a-446b-8c20-96dc0886b88f-openstack-config\") pod \"openstackclient\" (UID: \"eefa17c1-494a-446b-8c20-96dc0886b88f\") " pod="openstack/openstackclient" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.924414 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19a55666-374f-430a-a80b-f61538cbfb6b" path="/var/lib/kubelet/pods/19a55666-374f-430a-a80b-f61538cbfb6b/volumes" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.967892 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/eefa17c1-494a-446b-8c20-96dc0886b88f-openstack-config\") pod \"openstackclient\" (UID: \"eefa17c1-494a-446b-8c20-96dc0886b88f\") " pod="openstack/openstackclient" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.967990 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/eefa17c1-494a-446b-8c20-96dc0886b88f-openstack-config-secret\") pod \"openstackclient\" (UID: \"eefa17c1-494a-446b-8c20-96dc0886b88f\") " pod="openstack/openstackclient" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.968036 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfg7l\" (UniqueName: \"kubernetes.io/projected/eefa17c1-494a-446b-8c20-96dc0886b88f-kube-api-access-gfg7l\") pod \"openstackclient\" (UID: \"eefa17c1-494a-446b-8c20-96dc0886b88f\") " pod="openstack/openstackclient" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.968080 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eefa17c1-494a-446b-8c20-96dc0886b88f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"eefa17c1-494a-446b-8c20-96dc0886b88f\") " pod="openstack/openstackclient" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.969072 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/eefa17c1-494a-446b-8c20-96dc0886b88f-openstack-config\") pod \"openstackclient\" (UID: \"eefa17c1-494a-446b-8c20-96dc0886b88f\") " pod="openstack/openstackclient" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.973187 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/eefa17c1-494a-446b-8c20-96dc0886b88f-openstack-config-secret\") pod \"openstackclient\" (UID: \"eefa17c1-494a-446b-8c20-96dc0886b88f\") " pod="openstack/openstackclient" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.979305 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eefa17c1-494a-446b-8c20-96dc0886b88f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"eefa17c1-494a-446b-8c20-96dc0886b88f\") " pod="openstack/openstackclient" Oct 02 10:07:52 crc kubenswrapper[4934]: I1002 10:07:52.989422 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfg7l\" (UniqueName: \"kubernetes.io/projected/eefa17c1-494a-446b-8c20-96dc0886b88f-kube-api-access-gfg7l\") pod \"openstackclient\" (UID: \"eefa17c1-494a-446b-8c20-96dc0886b88f\") " pod="openstack/openstackclient" Oct 02 10:07:53 crc kubenswrapper[4934]: I1002 10:07:53.110664 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 02 10:07:53 crc kubenswrapper[4934]: I1002 10:07:53.111787 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 10:07:53 crc kubenswrapper[4934]: I1002 10:07:53.136764 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 02 10:07:53 crc kubenswrapper[4934]: I1002 10:07:53.160452 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 02 10:07:53 crc kubenswrapper[4934]: I1002 10:07:53.161650 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 10:07:53 crc kubenswrapper[4934]: I1002 10:07:53.186388 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 02 10:07:53 crc kubenswrapper[4934]: I1002 10:07:53.262718 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a6807b23-4962-4a52-8905-af632dbf9b5c","Type":"ContainerStarted","Data":"cf136300afd2954855f33d84c657f64620e35bd32b172a2de80850d520c46277"} Oct 02 10:07:53 crc kubenswrapper[4934]: I1002 10:07:53.263516 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a6807b23-4962-4a52-8905-af632dbf9b5c","Type":"ContainerStarted","Data":"5ab2b91762ee0df2e77c6a1f5fabf370779d0fac21eb66b99279e1a4ad76a03d"} Oct 02 10:07:53 crc kubenswrapper[4934]: I1002 10:07:53.273345 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c99e264-05b3-4cd7-91ae-cabb6d10466a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9c99e264-05b3-4cd7-91ae-cabb6d10466a\") " pod="openstack/openstackclient" Oct 02 10:07:53 crc kubenswrapper[4934]: I1002 10:07:53.273496 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9c99e264-05b3-4cd7-91ae-cabb6d10466a-openstack-config-secret\") pod \"openstackclient\" (UID: \"9c99e264-05b3-4cd7-91ae-cabb6d10466a\") " pod="openstack/openstackclient" Oct 02 10:07:53 crc kubenswrapper[4934]: I1002 10:07:53.273764 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9c99e264-05b3-4cd7-91ae-cabb6d10466a-openstack-config\") pod \"openstackclient\" (UID: \"9c99e264-05b3-4cd7-91ae-cabb6d10466a\") " pod="openstack/openstackclient" Oct 02 10:07:53 crc kubenswrapper[4934]: I1002 10:07:53.273891 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96nfv\" (UniqueName: \"kubernetes.io/projected/9c99e264-05b3-4cd7-91ae-cabb6d10466a-kube-api-access-96nfv\") pod \"openstackclient\" (UID: \"9c99e264-05b3-4cd7-91ae-cabb6d10466a\") " pod="openstack/openstackclient" Oct 02 10:07:53 crc kubenswrapper[4934]: I1002 10:07:53.284106 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e0d0eed2-217a-4dbc-906a-797c09b32081","Type":"ContainerStarted","Data":"b5bec96098ee1d10a714cbf2c1e87475985099d9a3c2f3f34b61fdccced4a950"} Oct 02 10:07:53 crc kubenswrapper[4934]: I1002 10:07:53.286892 4934 generic.go:334] "Generic (PLEG): container finished" podID="f55848b6-dd77-46d8-8366-4715042cdd5d" containerID="79c57afcd39582909d2f4e613b9d5c29d54192176c3f8f3aaa960f62fd72a9ee" exitCode=0 Oct 02 10:07:53 crc kubenswrapper[4934]: I1002 10:07:53.287253 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79fb47bfff-98d57" event={"ID":"f55848b6-dd77-46d8-8366-4715042cdd5d","Type":"ContainerDied","Data":"79c57afcd39582909d2f4e613b9d5c29d54192176c3f8f3aaa960f62fd72a9ee"} Oct 02 10:07:53 crc kubenswrapper[4934]: I1002 10:07:53.288034 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79fb47bfff-98d57" event={"ID":"f55848b6-dd77-46d8-8366-4715042cdd5d","Type":"ContainerStarted","Data":"18645ce4ec2ad743abcf2b1feee55c89fe461a48841967d4b1116a5aefbce090"} Oct 02 10:07:53 crc kubenswrapper[4934]: E1002 10:07:53.338547 4934 log.go:32] "RunPodSandbox from runtime service failed" err=< Oct 02 10:07:53 crc kubenswrapper[4934]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_eefa17c1-494a-446b-8c20-96dc0886b88f_0(05140e4855f6240328ef6a66e345810c7945fbc01852c9aed58f81a6e03eb24f): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"05140e4855f6240328ef6a66e345810c7945fbc01852c9aed58f81a6e03eb24f" Netns:"/var/run/netns/6d711cb1-8822-4d9f-a903-57e73589ccfb" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=05140e4855f6240328ef6a66e345810c7945fbc01852c9aed58f81a6e03eb24f;K8S_POD_UID=eefa17c1-494a-446b-8c20-96dc0886b88f" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/eefa17c1-494a-446b-8c20-96dc0886b88f]: expected pod UID "eefa17c1-494a-446b-8c20-96dc0886b88f" but got "9c99e264-05b3-4cd7-91ae-cabb6d10466a" from Kube API Oct 02 10:07:53 crc kubenswrapper[4934]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Oct 02 10:07:53 crc kubenswrapper[4934]: > Oct 02 10:07:53 crc kubenswrapper[4934]: E1002 10:07:53.338891 4934 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Oct 02 10:07:53 crc kubenswrapper[4934]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_eefa17c1-494a-446b-8c20-96dc0886b88f_0(05140e4855f6240328ef6a66e345810c7945fbc01852c9aed58f81a6e03eb24f): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"05140e4855f6240328ef6a66e345810c7945fbc01852c9aed58f81a6e03eb24f" Netns:"/var/run/netns/6d711cb1-8822-4d9f-a903-57e73589ccfb" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=05140e4855f6240328ef6a66e345810c7945fbc01852c9aed58f81a6e03eb24f;K8S_POD_UID=eefa17c1-494a-446b-8c20-96dc0886b88f" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/eefa17c1-494a-446b-8c20-96dc0886b88f]: expected pod UID "eefa17c1-494a-446b-8c20-96dc0886b88f" but got "9c99e264-05b3-4cd7-91ae-cabb6d10466a" from Kube API Oct 02 10:07:53 crc kubenswrapper[4934]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Oct 02 10:07:53 crc kubenswrapper[4934]: > pod="openstack/openstackclient" Oct 02 10:07:53 crc kubenswrapper[4934]: I1002 10:07:53.375237 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9c99e264-05b3-4cd7-91ae-cabb6d10466a-openstack-config\") pod \"openstackclient\" (UID: \"9c99e264-05b3-4cd7-91ae-cabb6d10466a\") " pod="openstack/openstackclient" Oct 02 10:07:53 crc kubenswrapper[4934]: I1002 10:07:53.375286 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96nfv\" (UniqueName: \"kubernetes.io/projected/9c99e264-05b3-4cd7-91ae-cabb6d10466a-kube-api-access-96nfv\") pod \"openstackclient\" (UID: \"9c99e264-05b3-4cd7-91ae-cabb6d10466a\") " pod="openstack/openstackclient" Oct 02 10:07:53 crc kubenswrapper[4934]: I1002 10:07:53.375342 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c99e264-05b3-4cd7-91ae-cabb6d10466a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9c99e264-05b3-4cd7-91ae-cabb6d10466a\") " pod="openstack/openstackclient" Oct 02 10:07:53 crc kubenswrapper[4934]: I1002 10:07:53.375375 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9c99e264-05b3-4cd7-91ae-cabb6d10466a-openstack-config-secret\") pod \"openstackclient\" (UID: \"9c99e264-05b3-4cd7-91ae-cabb6d10466a\") " pod="openstack/openstackclient" Oct 02 10:07:53 crc kubenswrapper[4934]: I1002 10:07:53.377376 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9c99e264-05b3-4cd7-91ae-cabb6d10466a-openstack-config\") pod \"openstackclient\" (UID: \"9c99e264-05b3-4cd7-91ae-cabb6d10466a\") " pod="openstack/openstackclient" Oct 02 10:07:53 crc kubenswrapper[4934]: I1002 10:07:53.382509 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c99e264-05b3-4cd7-91ae-cabb6d10466a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9c99e264-05b3-4cd7-91ae-cabb6d10466a\") " pod="openstack/openstackclient" Oct 02 10:07:53 crc kubenswrapper[4934]: I1002 10:07:53.385217 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9c99e264-05b3-4cd7-91ae-cabb6d10466a-openstack-config-secret\") pod \"openstackclient\" (UID: \"9c99e264-05b3-4cd7-91ae-cabb6d10466a\") " pod="openstack/openstackclient" Oct 02 10:07:53 crc kubenswrapper[4934]: I1002 10:07:53.392083 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96nfv\" (UniqueName: \"kubernetes.io/projected/9c99e264-05b3-4cd7-91ae-cabb6d10466a-kube-api-access-96nfv\") pod \"openstackclient\" (UID: \"9c99e264-05b3-4cd7-91ae-cabb6d10466a\") " pod="openstack/openstackclient" Oct 02 10:07:53 crc kubenswrapper[4934]: I1002 10:07:53.493004 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 10:07:54 crc kubenswrapper[4934]: I1002 10:07:54.017753 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 02 10:07:54 crc kubenswrapper[4934]: I1002 10:07:54.306005 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e0d0eed2-217a-4dbc-906a-797c09b32081","Type":"ContainerStarted","Data":"1bface884ab14709445ccfca87f0780807ad68ad0f60be0fc9cc86ff635fa242"} Oct 02 10:07:54 crc kubenswrapper[4934]: I1002 10:07:54.320704 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79fb47bfff-98d57" event={"ID":"f55848b6-dd77-46d8-8366-4715042cdd5d","Type":"ContainerStarted","Data":"df78f82edde3a7888cee3d1396b9ac2314053be6d07b967766fd37331acc4de7"} Oct 02 10:07:54 crc kubenswrapper[4934]: I1002 10:07:54.320763 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79fb47bfff-98d57" Oct 02 10:07:54 crc kubenswrapper[4934]: I1002 10:07:54.334835 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a6807b23-4962-4a52-8905-af632dbf9b5c","Type":"ContainerStarted","Data":"d3703465d95b0650d918396201608ca3e0f9a59a3224c6d8ed98cd8249f84cc2"} Oct 02 10:07:54 crc kubenswrapper[4934]: I1002 10:07:54.335821 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 02 10:07:54 crc kubenswrapper[4934]: I1002 10:07:54.345330 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79fb47bfff-98d57" podStartSLOduration=3.345315017 podStartE2EDuration="3.345315017s" podCreationTimestamp="2025-10-02 10:07:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:07:54.343706411 +0000 UTC m=+1146.096347933" watchObservedRunningTime="2025-10-02 10:07:54.345315017 +0000 UTC m=+1146.097956539" Oct 02 10:07:54 crc kubenswrapper[4934]: I1002 10:07:54.355090 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 10:07:54 crc kubenswrapper[4934]: I1002 10:07:54.355746 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"9c99e264-05b3-4cd7-91ae-cabb6d10466a","Type":"ContainerStarted","Data":"5f5a8ec06cf544ec99a01279987fc63735e5bb55e959d550ef31cc4e434b7c18"} Oct 02 10:07:54 crc kubenswrapper[4934]: I1002 10:07:54.369305 4934 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="eefa17c1-494a-446b-8c20-96dc0886b88f" podUID="9c99e264-05b3-4cd7-91ae-cabb6d10466a" Oct 02 10:07:54 crc kubenswrapper[4934]: I1002 10:07:54.376448 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.376425044 podStartE2EDuration="3.376425044s" podCreationTimestamp="2025-10-02 10:07:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:07:54.366534702 +0000 UTC m=+1146.119176234" watchObservedRunningTime="2025-10-02 10:07:54.376425044 +0000 UTC m=+1146.129066576" Oct 02 10:07:54 crc kubenswrapper[4934]: I1002 10:07:54.416613 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 10:07:54 crc kubenswrapper[4934]: I1002 10:07:54.443174 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 02 10:07:54 crc kubenswrapper[4934]: I1002 10:07:54.505221 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/eefa17c1-494a-446b-8c20-96dc0886b88f-openstack-config-secret\") pod \"eefa17c1-494a-446b-8c20-96dc0886b88f\" (UID: \"eefa17c1-494a-446b-8c20-96dc0886b88f\") " Oct 02 10:07:54 crc kubenswrapper[4934]: I1002 10:07:54.510667 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eefa17c1-494a-446b-8c20-96dc0886b88f-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "eefa17c1-494a-446b-8c20-96dc0886b88f" (UID: "eefa17c1-494a-446b-8c20-96dc0886b88f"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:07:54 crc kubenswrapper[4934]: I1002 10:07:54.607183 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfg7l\" (UniqueName: \"kubernetes.io/projected/eefa17c1-494a-446b-8c20-96dc0886b88f-kube-api-access-gfg7l\") pod \"eefa17c1-494a-446b-8c20-96dc0886b88f\" (UID: \"eefa17c1-494a-446b-8c20-96dc0886b88f\") " Oct 02 10:07:54 crc kubenswrapper[4934]: I1002 10:07:54.607266 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/eefa17c1-494a-446b-8c20-96dc0886b88f-openstack-config\") pod \"eefa17c1-494a-446b-8c20-96dc0886b88f\" (UID: \"eefa17c1-494a-446b-8c20-96dc0886b88f\") " Oct 02 10:07:54 crc kubenswrapper[4934]: I1002 10:07:54.607333 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eefa17c1-494a-446b-8c20-96dc0886b88f-combined-ca-bundle\") pod \"eefa17c1-494a-446b-8c20-96dc0886b88f\" (UID: \"eefa17c1-494a-446b-8c20-96dc0886b88f\") " Oct 02 10:07:54 crc kubenswrapper[4934]: I1002 10:07:54.607697 4934 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/eefa17c1-494a-446b-8c20-96dc0886b88f-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:54 crc kubenswrapper[4934]: I1002 10:07:54.607969 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eefa17c1-494a-446b-8c20-96dc0886b88f-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "eefa17c1-494a-446b-8c20-96dc0886b88f" (UID: "eefa17c1-494a-446b-8c20-96dc0886b88f"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:07:54 crc kubenswrapper[4934]: I1002 10:07:54.610423 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eefa17c1-494a-446b-8c20-96dc0886b88f-kube-api-access-gfg7l" (OuterVolumeSpecName: "kube-api-access-gfg7l") pod "eefa17c1-494a-446b-8c20-96dc0886b88f" (UID: "eefa17c1-494a-446b-8c20-96dc0886b88f"). InnerVolumeSpecName "kube-api-access-gfg7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:07:54 crc kubenswrapper[4934]: I1002 10:07:54.610847 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eefa17c1-494a-446b-8c20-96dc0886b88f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eefa17c1-494a-446b-8c20-96dc0886b88f" (UID: "eefa17c1-494a-446b-8c20-96dc0886b88f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:07:54 crc kubenswrapper[4934]: I1002 10:07:54.710112 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfg7l\" (UniqueName: \"kubernetes.io/projected/eefa17c1-494a-446b-8c20-96dc0886b88f-kube-api-access-gfg7l\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:54 crc kubenswrapper[4934]: I1002 10:07:54.710141 4934 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/eefa17c1-494a-446b-8c20-96dc0886b88f-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:54 crc kubenswrapper[4934]: I1002 10:07:54.710458 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eefa17c1-494a-446b-8c20-96dc0886b88f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:54 crc kubenswrapper[4934]: I1002 10:07:54.925717 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eefa17c1-494a-446b-8c20-96dc0886b88f" path="/var/lib/kubelet/pods/eefa17c1-494a-446b-8c20-96dc0886b88f/volumes" Oct 02 10:07:55 crc kubenswrapper[4934]: I1002 10:07:55.388808 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 10:07:55 crc kubenswrapper[4934]: I1002 10:07:55.391041 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e0d0eed2-217a-4dbc-906a-797c09b32081","Type":"ContainerStarted","Data":"8c229984e28e11867bf51914dc16c055d7e38ce173f5c33e9bf9b496ae9e06d5"} Oct 02 10:07:55 crc kubenswrapper[4934]: I1002 10:07:55.419706 4934 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="eefa17c1-494a-446b-8c20-96dc0886b88f" podUID="9c99e264-05b3-4cd7-91ae-cabb6d10466a" Oct 02 10:07:55 crc kubenswrapper[4934]: I1002 10:07:55.422813 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.557439643 podStartE2EDuration="4.422795698s" podCreationTimestamp="2025-10-02 10:07:51 +0000 UTC" firstStartedPulling="2025-10-02 10:07:52.34281191 +0000 UTC m=+1144.095453432" lastFinishedPulling="2025-10-02 10:07:53.208167965 +0000 UTC m=+1144.960809487" observedRunningTime="2025-10-02 10:07:55.416393336 +0000 UTC m=+1147.169034858" watchObservedRunningTime="2025-10-02 10:07:55.422795698 +0000 UTC m=+1147.175437220" Oct 02 10:07:56 crc kubenswrapper[4934]: I1002 10:07:56.397932 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="a6807b23-4962-4a52-8905-af632dbf9b5c" containerName="cinder-api-log" containerID="cri-o://cf136300afd2954855f33d84c657f64620e35bd32b172a2de80850d520c46277" gracePeriod=30 Oct 02 10:07:56 crc kubenswrapper[4934]: I1002 10:07:56.397991 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="a6807b23-4962-4a52-8905-af632dbf9b5c" containerName="cinder-api" containerID="cri-o://d3703465d95b0650d918396201608ca3e0f9a59a3224c6d8ed98cd8249f84cc2" gracePeriod=30 Oct 02 10:07:56 crc kubenswrapper[4934]: I1002 10:07:56.789018 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.074006 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.185371 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.270461 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6807b23-4962-4a52-8905-af632dbf9b5c-combined-ca-bundle\") pod \"a6807b23-4962-4a52-8905-af632dbf9b5c\" (UID: \"a6807b23-4962-4a52-8905-af632dbf9b5c\") " Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.270532 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6807b23-4962-4a52-8905-af632dbf9b5c-scripts\") pod \"a6807b23-4962-4a52-8905-af632dbf9b5c\" (UID: \"a6807b23-4962-4a52-8905-af632dbf9b5c\") " Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.270592 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a6807b23-4962-4a52-8905-af632dbf9b5c-config-data-custom\") pod \"a6807b23-4962-4a52-8905-af632dbf9b5c\" (UID: \"a6807b23-4962-4a52-8905-af632dbf9b5c\") " Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.270639 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdvdz\" (UniqueName: \"kubernetes.io/projected/a6807b23-4962-4a52-8905-af632dbf9b5c-kube-api-access-mdvdz\") pod \"a6807b23-4962-4a52-8905-af632dbf9b5c\" (UID: \"a6807b23-4962-4a52-8905-af632dbf9b5c\") " Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.270698 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a6807b23-4962-4a52-8905-af632dbf9b5c-etc-machine-id\") pod \"a6807b23-4962-4a52-8905-af632dbf9b5c\" (UID: \"a6807b23-4962-4a52-8905-af632dbf9b5c\") " Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.270740 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6807b23-4962-4a52-8905-af632dbf9b5c-config-data\") pod \"a6807b23-4962-4a52-8905-af632dbf9b5c\" (UID: \"a6807b23-4962-4a52-8905-af632dbf9b5c\") " Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.270815 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6807b23-4962-4a52-8905-af632dbf9b5c-logs\") pod \"a6807b23-4962-4a52-8905-af632dbf9b5c\" (UID: \"a6807b23-4962-4a52-8905-af632dbf9b5c\") " Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.271524 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6807b23-4962-4a52-8905-af632dbf9b5c-logs" (OuterVolumeSpecName: "logs") pod "a6807b23-4962-4a52-8905-af632dbf9b5c" (UID: "a6807b23-4962-4a52-8905-af632dbf9b5c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.272502 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a6807b23-4962-4a52-8905-af632dbf9b5c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a6807b23-4962-4a52-8905-af632dbf9b5c" (UID: "a6807b23-4962-4a52-8905-af632dbf9b5c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.279444 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6807b23-4962-4a52-8905-af632dbf9b5c-kube-api-access-mdvdz" (OuterVolumeSpecName: "kube-api-access-mdvdz") pod "a6807b23-4962-4a52-8905-af632dbf9b5c" (UID: "a6807b23-4962-4a52-8905-af632dbf9b5c"). InnerVolumeSpecName "kube-api-access-mdvdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.291380 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6807b23-4962-4a52-8905-af632dbf9b5c-scripts" (OuterVolumeSpecName: "scripts") pod "a6807b23-4962-4a52-8905-af632dbf9b5c" (UID: "a6807b23-4962-4a52-8905-af632dbf9b5c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.294742 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6807b23-4962-4a52-8905-af632dbf9b5c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a6807b23-4962-4a52-8905-af632dbf9b5c" (UID: "a6807b23-4962-4a52-8905-af632dbf9b5c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.311706 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6807b23-4962-4a52-8905-af632dbf9b5c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a6807b23-4962-4a52-8905-af632dbf9b5c" (UID: "a6807b23-4962-4a52-8905-af632dbf9b5c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.333070 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6807b23-4962-4a52-8905-af632dbf9b5c-config-data" (OuterVolumeSpecName: "config-data") pod "a6807b23-4962-4a52-8905-af632dbf9b5c" (UID: "a6807b23-4962-4a52-8905-af632dbf9b5c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.374279 4934 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a6807b23-4962-4a52-8905-af632dbf9b5c-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.374321 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdvdz\" (UniqueName: \"kubernetes.io/projected/a6807b23-4962-4a52-8905-af632dbf9b5c-kube-api-access-mdvdz\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.374339 4934 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a6807b23-4962-4a52-8905-af632dbf9b5c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.374352 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6807b23-4962-4a52-8905-af632dbf9b5c-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.374369 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6807b23-4962-4a52-8905-af632dbf9b5c-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.374381 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6807b23-4962-4a52-8905-af632dbf9b5c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.374392 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6807b23-4962-4a52-8905-af632dbf9b5c-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.411439 4934 generic.go:334] "Generic (PLEG): container finished" podID="a6807b23-4962-4a52-8905-af632dbf9b5c" containerID="d3703465d95b0650d918396201608ca3e0f9a59a3224c6d8ed98cd8249f84cc2" exitCode=0 Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.411467 4934 generic.go:334] "Generic (PLEG): container finished" podID="a6807b23-4962-4a52-8905-af632dbf9b5c" containerID="cf136300afd2954855f33d84c657f64620e35bd32b172a2de80850d520c46277" exitCode=143 Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.411517 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.411555 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a6807b23-4962-4a52-8905-af632dbf9b5c","Type":"ContainerDied","Data":"d3703465d95b0650d918396201608ca3e0f9a59a3224c6d8ed98cd8249f84cc2"} Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.411621 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a6807b23-4962-4a52-8905-af632dbf9b5c","Type":"ContainerDied","Data":"cf136300afd2954855f33d84c657f64620e35bd32b172a2de80850d520c46277"} Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.411638 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a6807b23-4962-4a52-8905-af632dbf9b5c","Type":"ContainerDied","Data":"5ab2b91762ee0df2e77c6a1f5fabf370779d0fac21eb66b99279e1a4ad76a03d"} Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.411679 4934 scope.go:117] "RemoveContainer" containerID="d3703465d95b0650d918396201608ca3e0f9a59a3224c6d8ed98cd8249f84cc2" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.445629 4934 scope.go:117] "RemoveContainer" containerID="cf136300afd2954855f33d84c657f64620e35bd32b172a2de80850d520c46277" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.457607 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.484045 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.502321 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 02 10:07:57 crc kubenswrapper[4934]: E1002 10:07:57.502740 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6807b23-4962-4a52-8905-af632dbf9b5c" containerName="cinder-api" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.502757 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6807b23-4962-4a52-8905-af632dbf9b5c" containerName="cinder-api" Oct 02 10:07:57 crc kubenswrapper[4934]: E1002 10:07:57.502771 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6807b23-4962-4a52-8905-af632dbf9b5c" containerName="cinder-api-log" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.502778 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6807b23-4962-4a52-8905-af632dbf9b5c" containerName="cinder-api-log" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.502971 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6807b23-4962-4a52-8905-af632dbf9b5c" containerName="cinder-api-log" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.502998 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6807b23-4962-4a52-8905-af632dbf9b5c" containerName="cinder-api" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.503897 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.506067 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.506256 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.506443 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.508623 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.553754 4934 scope.go:117] "RemoveContainer" containerID="d3703465d95b0650d918396201608ca3e0f9a59a3224c6d8ed98cd8249f84cc2" Oct 02 10:07:57 crc kubenswrapper[4934]: E1002 10:07:57.558747 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3703465d95b0650d918396201608ca3e0f9a59a3224c6d8ed98cd8249f84cc2\": container with ID starting with d3703465d95b0650d918396201608ca3e0f9a59a3224c6d8ed98cd8249f84cc2 not found: ID does not exist" containerID="d3703465d95b0650d918396201608ca3e0f9a59a3224c6d8ed98cd8249f84cc2" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.558791 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3703465d95b0650d918396201608ca3e0f9a59a3224c6d8ed98cd8249f84cc2"} err="failed to get container status \"d3703465d95b0650d918396201608ca3e0f9a59a3224c6d8ed98cd8249f84cc2\": rpc error: code = NotFound desc = could not find container \"d3703465d95b0650d918396201608ca3e0f9a59a3224c6d8ed98cd8249f84cc2\": container with ID starting with d3703465d95b0650d918396201608ca3e0f9a59a3224c6d8ed98cd8249f84cc2 not found: ID does not exist" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.558816 4934 scope.go:117] "RemoveContainer" containerID="cf136300afd2954855f33d84c657f64620e35bd32b172a2de80850d520c46277" Oct 02 10:07:57 crc kubenswrapper[4934]: E1002 10:07:57.559308 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf136300afd2954855f33d84c657f64620e35bd32b172a2de80850d520c46277\": container with ID starting with cf136300afd2954855f33d84c657f64620e35bd32b172a2de80850d520c46277 not found: ID does not exist" containerID="cf136300afd2954855f33d84c657f64620e35bd32b172a2de80850d520c46277" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.559338 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf136300afd2954855f33d84c657f64620e35bd32b172a2de80850d520c46277"} err="failed to get container status \"cf136300afd2954855f33d84c657f64620e35bd32b172a2de80850d520c46277\": rpc error: code = NotFound desc = could not find container \"cf136300afd2954855f33d84c657f64620e35bd32b172a2de80850d520c46277\": container with ID starting with cf136300afd2954855f33d84c657f64620e35bd32b172a2de80850d520c46277 not found: ID does not exist" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.559377 4934 scope.go:117] "RemoveContainer" containerID="d3703465d95b0650d918396201608ca3e0f9a59a3224c6d8ed98cd8249f84cc2" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.560282 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3703465d95b0650d918396201608ca3e0f9a59a3224c6d8ed98cd8249f84cc2"} err="failed to get container status \"d3703465d95b0650d918396201608ca3e0f9a59a3224c6d8ed98cd8249f84cc2\": rpc error: code = NotFound desc = could not find container \"d3703465d95b0650d918396201608ca3e0f9a59a3224c6d8ed98cd8249f84cc2\": container with ID starting with d3703465d95b0650d918396201608ca3e0f9a59a3224c6d8ed98cd8249f84cc2 not found: ID does not exist" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.560308 4934 scope.go:117] "RemoveContainer" containerID="cf136300afd2954855f33d84c657f64620e35bd32b172a2de80850d520c46277" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.560639 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf136300afd2954855f33d84c657f64620e35bd32b172a2de80850d520c46277"} err="failed to get container status \"cf136300afd2954855f33d84c657f64620e35bd32b172a2de80850d520c46277\": rpc error: code = NotFound desc = could not find container \"cf136300afd2954855f33d84c657f64620e35bd32b172a2de80850d520c46277\": container with ID starting with cf136300afd2954855f33d84c657f64620e35bd32b172a2de80850d520c46277 not found: ID does not exist" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.678708 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-config-data\") pod \"cinder-api-0\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.678752 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd41dc17-6659-4b43-9758-f67d831993ba-logs\") pod \"cinder-api-0\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.678792 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.679047 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.679136 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cd41dc17-6659-4b43-9758-f67d831993ba-etc-machine-id\") pod \"cinder-api-0\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.679202 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-config-data-custom\") pod \"cinder-api-0\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.679449 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-public-tls-certs\") pod \"cinder-api-0\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.679484 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sf2w\" (UniqueName: \"kubernetes.io/projected/cd41dc17-6659-4b43-9758-f67d831993ba-kube-api-access-7sf2w\") pod \"cinder-api-0\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.679556 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-scripts\") pod \"cinder-api-0\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.781747 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-public-tls-certs\") pod \"cinder-api-0\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.781789 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sf2w\" (UniqueName: \"kubernetes.io/projected/cd41dc17-6659-4b43-9758-f67d831993ba-kube-api-access-7sf2w\") pod \"cinder-api-0\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.781824 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-scripts\") pod \"cinder-api-0\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.781850 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-config-data\") pod \"cinder-api-0\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.781865 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd41dc17-6659-4b43-9758-f67d831993ba-logs\") pod \"cinder-api-0\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.781893 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.781927 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.781943 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cd41dc17-6659-4b43-9758-f67d831993ba-etc-machine-id\") pod \"cinder-api-0\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.781960 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-config-data-custom\") pod \"cinder-api-0\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.782569 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cd41dc17-6659-4b43-9758-f67d831993ba-etc-machine-id\") pod \"cinder-api-0\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.782857 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd41dc17-6659-4b43-9758-f67d831993ba-logs\") pod \"cinder-api-0\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.787443 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-config-data-custom\") pod \"cinder-api-0\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.787941 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.789350 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-scripts\") pod \"cinder-api-0\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.790394 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-public-tls-certs\") pod \"cinder-api-0\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.797893 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.801962 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-config-data\") pod \"cinder-api-0\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.803061 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sf2w\" (UniqueName: \"kubernetes.io/projected/cd41dc17-6659-4b43-9758-f67d831993ba-kube-api-access-7sf2w\") pod \"cinder-api-0\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " pod="openstack/cinder-api-0" Oct 02 10:07:57 crc kubenswrapper[4934]: I1002 10:07:57.877019 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 10:07:58 crc kubenswrapper[4934]: I1002 10:07:58.367971 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 10:07:58 crc kubenswrapper[4934]: W1002 10:07:58.377712 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd41dc17_6659_4b43_9758_f67d831993ba.slice/crio-e1c2d28a7763a945c2fdf6581a667b801498dd55a55f90a0c565eaf3b399e221 WatchSource:0}: Error finding container e1c2d28a7763a945c2fdf6581a667b801498dd55a55f90a0c565eaf3b399e221: Status 404 returned error can't find the container with id e1c2d28a7763a945c2fdf6581a667b801498dd55a55f90a0c565eaf3b399e221 Oct 02 10:07:58 crc kubenswrapper[4934]: I1002 10:07:58.425493 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cd41dc17-6659-4b43-9758-f67d831993ba","Type":"ContainerStarted","Data":"e1c2d28a7763a945c2fdf6581a667b801498dd55a55f90a0c565eaf3b399e221"} Oct 02 10:07:58 crc kubenswrapper[4934]: I1002 10:07:58.736024 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-57c45ff45d-ghdvr" Oct 02 10:07:58 crc kubenswrapper[4934]: I1002 10:07:58.848469 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-57c45ff45d-ghdvr" Oct 02 10:07:58 crc kubenswrapper[4934]: I1002 10:07:58.949338 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6807b23-4962-4a52-8905-af632dbf9b5c" path="/var/lib/kubelet/pods/a6807b23-4962-4a52-8905-af632dbf9b5c/volumes" Oct 02 10:07:58 crc kubenswrapper[4934]: I1002 10:07:58.950613 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:07:59 crc kubenswrapper[4934]: I1002 10:07:59.029237 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-57c45ff45d-ghdvr"] Oct 02 10:07:59 crc kubenswrapper[4934]: I1002 10:07:59.440941 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cd41dc17-6659-4b43-9758-f67d831993ba","Type":"ContainerStarted","Data":"ddc003d07d24a576f93d547d9f6f180b883faaf22fef13259bf076d508867ebe"} Oct 02 10:07:59 crc kubenswrapper[4934]: I1002 10:07:59.743402 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:07:59 crc kubenswrapper[4934]: I1002 10:07:59.743755 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="edb76b11-9129-495d-b44d-e998c3e8dceb" containerName="ceilometer-central-agent" containerID="cri-o://3e597152c8970caf6df4198a19c0fb1ca0108eab7f0f4e07d15ecd0797056ed0" gracePeriod=30 Oct 02 10:07:59 crc kubenswrapper[4934]: I1002 10:07:59.744714 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="edb76b11-9129-495d-b44d-e998c3e8dceb" containerName="proxy-httpd" containerID="cri-o://07a3ebc27c86cc0989e83071c71d3c33a754a7f416d67889d187120a9dc11c03" gracePeriod=30 Oct 02 10:07:59 crc kubenswrapper[4934]: I1002 10:07:59.745058 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="edb76b11-9129-495d-b44d-e998c3e8dceb" containerName="ceilometer-notification-agent" containerID="cri-o://cee5b6bd1bcc9c2981f3668a02ad37893a0ce6cb09bbaf102043a1606140b34f" gracePeriod=30 Oct 02 10:07:59 crc kubenswrapper[4934]: I1002 10:07:59.745100 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="edb76b11-9129-495d-b44d-e998c3e8dceb" containerName="sg-core" containerID="cri-o://3ae22becac0e4812e95a08cb7d2bbcf1e151ec55f7db552c2af6ec857bd58070" gracePeriod=30 Oct 02 10:07:59 crc kubenswrapper[4934]: I1002 10:07:59.767530 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.202765 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-bfc5bc4b9-crmsr"] Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.205409 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.210588 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.211062 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.211201 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.212829 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-bfc5bc4b9-crmsr"] Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.252501 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcxvb\" (UniqueName: \"kubernetes.io/projected/a30daf4c-06f2-48f5-b57f-f6704fae13a9-kube-api-access-xcxvb\") pod \"swift-proxy-bfc5bc4b9-crmsr\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.252615 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a30daf4c-06f2-48f5-b57f-f6704fae13a9-etc-swift\") pod \"swift-proxy-bfc5bc4b9-crmsr\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.252645 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a30daf4c-06f2-48f5-b57f-f6704fae13a9-run-httpd\") pod \"swift-proxy-bfc5bc4b9-crmsr\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.252661 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a30daf4c-06f2-48f5-b57f-f6704fae13a9-public-tls-certs\") pod \"swift-proxy-bfc5bc4b9-crmsr\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.252717 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a30daf4c-06f2-48f5-b57f-f6704fae13a9-config-data\") pod \"swift-proxy-bfc5bc4b9-crmsr\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.252734 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a30daf4c-06f2-48f5-b57f-f6704fae13a9-combined-ca-bundle\") pod \"swift-proxy-bfc5bc4b9-crmsr\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.252768 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a30daf4c-06f2-48f5-b57f-f6704fae13a9-log-httpd\") pod \"swift-proxy-bfc5bc4b9-crmsr\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.252824 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a30daf4c-06f2-48f5-b57f-f6704fae13a9-internal-tls-certs\") pod \"swift-proxy-bfc5bc4b9-crmsr\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.354258 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a30daf4c-06f2-48f5-b57f-f6704fae13a9-internal-tls-certs\") pod \"swift-proxy-bfc5bc4b9-crmsr\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.354359 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcxvb\" (UniqueName: \"kubernetes.io/projected/a30daf4c-06f2-48f5-b57f-f6704fae13a9-kube-api-access-xcxvb\") pod \"swift-proxy-bfc5bc4b9-crmsr\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.354398 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a30daf4c-06f2-48f5-b57f-f6704fae13a9-etc-swift\") pod \"swift-proxy-bfc5bc4b9-crmsr\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.354425 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a30daf4c-06f2-48f5-b57f-f6704fae13a9-run-httpd\") pod \"swift-proxy-bfc5bc4b9-crmsr\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.354442 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a30daf4c-06f2-48f5-b57f-f6704fae13a9-public-tls-certs\") pod \"swift-proxy-bfc5bc4b9-crmsr\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.354489 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a30daf4c-06f2-48f5-b57f-f6704fae13a9-config-data\") pod \"swift-proxy-bfc5bc4b9-crmsr\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.354510 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a30daf4c-06f2-48f5-b57f-f6704fae13a9-combined-ca-bundle\") pod \"swift-proxy-bfc5bc4b9-crmsr\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.354532 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a30daf4c-06f2-48f5-b57f-f6704fae13a9-log-httpd\") pod \"swift-proxy-bfc5bc4b9-crmsr\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.355026 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a30daf4c-06f2-48f5-b57f-f6704fae13a9-run-httpd\") pod \"swift-proxy-bfc5bc4b9-crmsr\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.355107 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a30daf4c-06f2-48f5-b57f-f6704fae13a9-log-httpd\") pod \"swift-proxy-bfc5bc4b9-crmsr\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.361978 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a30daf4c-06f2-48f5-b57f-f6704fae13a9-combined-ca-bundle\") pod \"swift-proxy-bfc5bc4b9-crmsr\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.362523 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a30daf4c-06f2-48f5-b57f-f6704fae13a9-public-tls-certs\") pod \"swift-proxy-bfc5bc4b9-crmsr\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.362632 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a30daf4c-06f2-48f5-b57f-f6704fae13a9-etc-swift\") pod \"swift-proxy-bfc5bc4b9-crmsr\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.362762 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a30daf4c-06f2-48f5-b57f-f6704fae13a9-internal-tls-certs\") pod \"swift-proxy-bfc5bc4b9-crmsr\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.365214 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a30daf4c-06f2-48f5-b57f-f6704fae13a9-config-data\") pod \"swift-proxy-bfc5bc4b9-crmsr\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.377279 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcxvb\" (UniqueName: \"kubernetes.io/projected/a30daf4c-06f2-48f5-b57f-f6704fae13a9-kube-api-access-xcxvb\") pod \"swift-proxy-bfc5bc4b9-crmsr\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.465497 4934 generic.go:334] "Generic (PLEG): container finished" podID="edb76b11-9129-495d-b44d-e998c3e8dceb" containerID="07a3ebc27c86cc0989e83071c71d3c33a754a7f416d67889d187120a9dc11c03" exitCode=0 Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.465526 4934 generic.go:334] "Generic (PLEG): container finished" podID="edb76b11-9129-495d-b44d-e998c3e8dceb" containerID="3ae22becac0e4812e95a08cb7d2bbcf1e151ec55f7db552c2af6ec857bd58070" exitCode=2 Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.465535 4934 generic.go:334] "Generic (PLEG): container finished" podID="edb76b11-9129-495d-b44d-e998c3e8dceb" containerID="3e597152c8970caf6df4198a19c0fb1ca0108eab7f0f4e07d15ecd0797056ed0" exitCode=0 Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.465595 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"edb76b11-9129-495d-b44d-e998c3e8dceb","Type":"ContainerDied","Data":"07a3ebc27c86cc0989e83071c71d3c33a754a7f416d67889d187120a9dc11c03"} Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.465622 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"edb76b11-9129-495d-b44d-e998c3e8dceb","Type":"ContainerDied","Data":"3ae22becac0e4812e95a08cb7d2bbcf1e151ec55f7db552c2af6ec857bd58070"} Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.465631 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"edb76b11-9129-495d-b44d-e998c3e8dceb","Type":"ContainerDied","Data":"3e597152c8970caf6df4198a19c0fb1ca0108eab7f0f4e07d15ecd0797056ed0"} Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.469016 4934 generic.go:334] "Generic (PLEG): container finished" podID="6427286d-a1ca-43ca-8ecb-bbdda45e360e" containerID="d4c05b6d8b3d3381710c9dd68250fc8af41ac47622cf6da8cc3795965ae20a68" exitCode=0 Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.469077 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jwpkw" event={"ID":"6427286d-a1ca-43ca-8ecb-bbdda45e360e","Type":"ContainerDied","Data":"d4c05b6d8b3d3381710c9dd68250fc8af41ac47622cf6da8cc3795965ae20a68"} Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.472168 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-57c45ff45d-ghdvr" podUID="a8799f51-9cec-454b-bdbe-860ce455baca" containerName="barbican-api-log" containerID="cri-o://ee4233b97847dce506924805dd7b618d62016dea5d61be04baec7613cd33b597" gracePeriod=30 Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.473056 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cd41dc17-6659-4b43-9758-f67d831993ba","Type":"ContainerStarted","Data":"86f78dcebecc392bfd4c80af3143e2df0ad38fe341e51b1a9c1ed70cc6f77bb2"} Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.473115 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.473150 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-57c45ff45d-ghdvr" podUID="a8799f51-9cec-454b-bdbe-860ce455baca" containerName="barbican-api" containerID="cri-o://59ff67fae05185b501e4f0db278a33b889fc417ce780a16ef043c8e691f5fcbd" gracePeriod=30 Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.484157 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-57c45ff45d-ghdvr" podUID="a8799f51-9cec-454b-bdbe-860ce455baca" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.151:9311/healthcheck\": EOF" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.484198 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-57c45ff45d-ghdvr" podUID="a8799f51-9cec-454b-bdbe-860ce455baca" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.151:9311/healthcheck\": EOF" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.534372 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:00 crc kubenswrapper[4934]: I1002 10:08:00.557216 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.557192504 podStartE2EDuration="3.557192504s" podCreationTimestamp="2025-10-02 10:07:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:08:00.535492905 +0000 UTC m=+1152.288134437" watchObservedRunningTime="2025-10-02 10:08:00.557192504 +0000 UTC m=+1152.309834026" Oct 02 10:08:01 crc kubenswrapper[4934]: I1002 10:08:01.485623 4934 generic.go:334] "Generic (PLEG): container finished" podID="edb76b11-9129-495d-b44d-e998c3e8dceb" containerID="cee5b6bd1bcc9c2981f3668a02ad37893a0ce6cb09bbaf102043a1606140b34f" exitCode=0 Oct 02 10:08:01 crc kubenswrapper[4934]: I1002 10:08:01.485684 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"edb76b11-9129-495d-b44d-e998c3e8dceb","Type":"ContainerDied","Data":"cee5b6bd1bcc9c2981f3668a02ad37893a0ce6cb09bbaf102043a1606140b34f"} Oct 02 10:08:01 crc kubenswrapper[4934]: I1002 10:08:01.498444 4934 generic.go:334] "Generic (PLEG): container finished" podID="a8799f51-9cec-454b-bdbe-860ce455baca" containerID="ee4233b97847dce506924805dd7b618d62016dea5d61be04baec7613cd33b597" exitCode=143 Oct 02 10:08:01 crc kubenswrapper[4934]: I1002 10:08:01.498710 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57c45ff45d-ghdvr" event={"ID":"a8799f51-9cec-454b-bdbe-860ce455baca","Type":"ContainerDied","Data":"ee4233b97847dce506924805dd7b618d62016dea5d61be04baec7613cd33b597"} Oct 02 10:08:01 crc kubenswrapper[4934]: I1002 10:08:01.933010 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79fb47bfff-98d57" Oct 02 10:08:02 crc kubenswrapper[4934]: I1002 10:08:02.023163 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-685744b6b9-5mdh2"] Oct 02 10:08:02 crc kubenswrapper[4934]: I1002 10:08:02.023412 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" podUID="f2059861-459c-448a-933d-bc0d7660a9dc" containerName="dnsmasq-dns" containerID="cri-o://a2aa6bf2a4073de25d3a5fb820f9b97b8128dc4b37142dd94163513f5caff7bc" gracePeriod=10 Oct 02 10:08:02 crc kubenswrapper[4934]: I1002 10:08:02.511727 4934 generic.go:334] "Generic (PLEG): container finished" podID="f2059861-459c-448a-933d-bc0d7660a9dc" containerID="a2aa6bf2a4073de25d3a5fb820f9b97b8128dc4b37142dd94163513f5caff7bc" exitCode=0 Oct 02 10:08:02 crc kubenswrapper[4934]: I1002 10:08:02.512156 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" event={"ID":"f2059861-459c-448a-933d-bc0d7660a9dc","Type":"ContainerDied","Data":"a2aa6bf2a4073de25d3a5fb820f9b97b8128dc4b37142dd94163513f5caff7bc"} Oct 02 10:08:03 crc kubenswrapper[4934]: I1002 10:08:03.004786 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 02 10:08:03 crc kubenswrapper[4934]: I1002 10:08:03.078160 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 10:08:03 crc kubenswrapper[4934]: I1002 10:08:03.339796 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" podUID="f2059861-459c-448a-933d-bc0d7660a9dc" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.142:5353: connect: connection refused" Oct 02 10:08:03 crc kubenswrapper[4934]: I1002 10:08:03.521473 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="e0d0eed2-217a-4dbc-906a-797c09b32081" containerName="cinder-scheduler" containerID="cri-o://1bface884ab14709445ccfca87f0780807ad68ad0f60be0fc9cc86ff635fa242" gracePeriod=30 Oct 02 10:08:03 crc kubenswrapper[4934]: I1002 10:08:03.521632 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="e0d0eed2-217a-4dbc-906a-797c09b32081" containerName="probe" containerID="cri-o://8c229984e28e11867bf51914dc16c055d7e38ce173f5c33e9bf9b496ae9e06d5" gracePeriod=30 Oct 02 10:08:04 crc kubenswrapper[4934]: I1002 10:08:04.268768 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-8648b56f4-5f8xk" podUID="350928d4-5a0b-4331-b1c2-220517d313a5" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.152:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 10:08:04 crc kubenswrapper[4934]: I1002 10:08:04.777508 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-bfc5bc4b9-crmsr"] Oct 02 10:08:05 crc kubenswrapper[4934]: I1002 10:08:05.541568 4934 generic.go:334] "Generic (PLEG): container finished" podID="e0d0eed2-217a-4dbc-906a-797c09b32081" containerID="8c229984e28e11867bf51914dc16c055d7e38ce173f5c33e9bf9b496ae9e06d5" exitCode=0 Oct 02 10:08:05 crc kubenswrapper[4934]: I1002 10:08:05.541960 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e0d0eed2-217a-4dbc-906a-797c09b32081","Type":"ContainerDied","Data":"8c229984e28e11867bf51914dc16c055d7e38ce173f5c33e9bf9b496ae9e06d5"} Oct 02 10:08:05 crc kubenswrapper[4934]: I1002 10:08:05.902440 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-57c45ff45d-ghdvr" podUID="a8799f51-9cec-454b-bdbe-860ce455baca" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.151:9311/healthcheck\": read tcp 10.217.0.2:55102->10.217.0.151:9311: read: connection reset by peer" Oct 02 10:08:05 crc kubenswrapper[4934]: I1002 10:08:05.902501 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-57c45ff45d-ghdvr" podUID="a8799f51-9cec-454b-bdbe-860ce455baca" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.151:9311/healthcheck\": read tcp 10.217.0.2:55116->10.217.0.151:9311: read: connection reset by peer" Oct 02 10:08:06 crc kubenswrapper[4934]: I1002 10:08:06.555213 4934 generic.go:334] "Generic (PLEG): container finished" podID="a8799f51-9cec-454b-bdbe-860ce455baca" containerID="59ff67fae05185b501e4f0db278a33b889fc417ce780a16ef043c8e691f5fcbd" exitCode=0 Oct 02 10:08:06 crc kubenswrapper[4934]: I1002 10:08:06.555400 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57c45ff45d-ghdvr" event={"ID":"a8799f51-9cec-454b-bdbe-860ce455baca","Type":"ContainerDied","Data":"59ff67fae05185b501e4f0db278a33b889fc417ce780a16ef043c8e691f5fcbd"} Oct 02 10:08:06 crc kubenswrapper[4934]: I1002 10:08:06.558250 4934 generic.go:334] "Generic (PLEG): container finished" podID="e0d0eed2-217a-4dbc-906a-797c09b32081" containerID="1bface884ab14709445ccfca87f0780807ad68ad0f60be0fc9cc86ff635fa242" exitCode=0 Oct 02 10:08:06 crc kubenswrapper[4934]: I1002 10:08:06.558275 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e0d0eed2-217a-4dbc-906a-797c09b32081","Type":"ContainerDied","Data":"1bface884ab14709445ccfca87f0780807ad68ad0f60be0fc9cc86ff635fa242"} Oct 02 10:08:06 crc kubenswrapper[4934]: I1002 10:08:06.904490 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-57c45ff45d-ghdvr" podUID="a8799f51-9cec-454b-bdbe-860ce455baca" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.151:9311/healthcheck\": dial tcp 10.217.0.151:9311: connect: connection refused" Oct 02 10:08:06 crc kubenswrapper[4934]: I1002 10:08:06.904548 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-57c45ff45d-ghdvr" podUID="a8799f51-9cec-454b-bdbe-860ce455baca" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.151:9311/healthcheck\": dial tcp 10.217.0.151:9311: connect: connection refused" Oct 02 10:08:06 crc kubenswrapper[4934]: I1002 10:08:06.904600 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-57c45ff45d-ghdvr" Oct 02 10:08:07 crc kubenswrapper[4934]: W1002 10:08:07.412773 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda30daf4c_06f2_48f5_b57f_f6704fae13a9.slice/crio-83cc72498710eda3ccb7563ee5eb49f8455310af6525b949f517589cb3b7027e WatchSource:0}: Error finding container 83cc72498710eda3ccb7563ee5eb49f8455310af6525b949f517589cb3b7027e: Status 404 returned error can't find the container with id 83cc72498710eda3ccb7563ee5eb49f8455310af6525b949f517589cb3b7027e Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.579295 4934 generic.go:334] "Generic (PLEG): container finished" podID="3c03e58b-5a31-4295-ba05-d219daff78b1" containerID="fd0b5910cc8779a765c024bbeafee2d307597efe1d65c9d3dbff30dc96cca647" exitCode=0 Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.579674 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-9z9t6" event={"ID":"3c03e58b-5a31-4295-ba05-d219daff78b1","Type":"ContainerDied","Data":"fd0b5910cc8779a765c024bbeafee2d307597efe1d65c9d3dbff30dc96cca647"} Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.590909 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jwpkw" event={"ID":"6427286d-a1ca-43ca-8ecb-bbdda45e360e","Type":"ContainerDied","Data":"35c7791c313ed28fdfce6e3b778e7be233cc1a891db6dfa945b93e0eb921a6a6"} Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.590978 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35c7791c313ed28fdfce6e3b778e7be233cc1a891db6dfa945b93e0eb921a6a6" Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.593721 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-bfc5bc4b9-crmsr" event={"ID":"a30daf4c-06f2-48f5-b57f-f6704fae13a9","Type":"ContainerStarted","Data":"83cc72498710eda3ccb7563ee5eb49f8455310af6525b949f517589cb3b7027e"} Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.611043 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"edb76b11-9129-495d-b44d-e998c3e8dceb","Type":"ContainerDied","Data":"1997b7563b3c0b444d269d563331d094f5e6ccb2898a0f7c1adccc61176a2e00"} Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.611099 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1997b7563b3c0b444d269d563331d094f5e6ccb2898a0f7c1adccc61176a2e00" Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.659753 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.663387 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jwpkw" Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.668430 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edb76b11-9129-495d-b44d-e998c3e8dceb-combined-ca-bundle\") pod \"edb76b11-9129-495d-b44d-e998c3e8dceb\" (UID: \"edb76b11-9129-495d-b44d-e998c3e8dceb\") " Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.668532 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6427286d-a1ca-43ca-8ecb-bbdda45e360e-combined-ca-bundle\") pod \"6427286d-a1ca-43ca-8ecb-bbdda45e360e\" (UID: \"6427286d-a1ca-43ca-8ecb-bbdda45e360e\") " Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.668616 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edb76b11-9129-495d-b44d-e998c3e8dceb-scripts\") pod \"edb76b11-9129-495d-b44d-e998c3e8dceb\" (UID: \"edb76b11-9129-495d-b44d-e998c3e8dceb\") " Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.668668 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6427286d-a1ca-43ca-8ecb-bbdda45e360e-config-data\") pod \"6427286d-a1ca-43ca-8ecb-bbdda45e360e\" (UID: \"6427286d-a1ca-43ca-8ecb-bbdda45e360e\") " Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.668693 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edb76b11-9129-495d-b44d-e998c3e8dceb-config-data\") pod \"edb76b11-9129-495d-b44d-e998c3e8dceb\" (UID: \"edb76b11-9129-495d-b44d-e998c3e8dceb\") " Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.668767 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edb76b11-9129-495d-b44d-e998c3e8dceb-log-httpd\") pod \"edb76b11-9129-495d-b44d-e998c3e8dceb\" (UID: \"edb76b11-9129-495d-b44d-e998c3e8dceb\") " Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.668808 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/edb76b11-9129-495d-b44d-e998c3e8dceb-sg-core-conf-yaml\") pod \"edb76b11-9129-495d-b44d-e998c3e8dceb\" (UID: \"edb76b11-9129-495d-b44d-e998c3e8dceb\") " Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.668832 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqq9t\" (UniqueName: \"kubernetes.io/projected/6427286d-a1ca-43ca-8ecb-bbdda45e360e-kube-api-access-zqq9t\") pod \"6427286d-a1ca-43ca-8ecb-bbdda45e360e\" (UID: \"6427286d-a1ca-43ca-8ecb-bbdda45e360e\") " Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.668853 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6427286d-a1ca-43ca-8ecb-bbdda45e360e-db-sync-config-data\") pod \"6427286d-a1ca-43ca-8ecb-bbdda45e360e\" (UID: \"6427286d-a1ca-43ca-8ecb-bbdda45e360e\") " Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.668869 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edb76b11-9129-495d-b44d-e998c3e8dceb-run-httpd\") pod \"edb76b11-9129-495d-b44d-e998c3e8dceb\" (UID: \"edb76b11-9129-495d-b44d-e998c3e8dceb\") " Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.668883 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79x2b\" (UniqueName: \"kubernetes.io/projected/edb76b11-9129-495d-b44d-e998c3e8dceb-kube-api-access-79x2b\") pod \"edb76b11-9129-495d-b44d-e998c3e8dceb\" (UID: \"edb76b11-9129-495d-b44d-e998c3e8dceb\") " Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.669973 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edb76b11-9129-495d-b44d-e998c3e8dceb-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "edb76b11-9129-495d-b44d-e998c3e8dceb" (UID: "edb76b11-9129-495d-b44d-e998c3e8dceb"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.670665 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edb76b11-9129-495d-b44d-e998c3e8dceb-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "edb76b11-9129-495d-b44d-e998c3e8dceb" (UID: "edb76b11-9129-495d-b44d-e998c3e8dceb"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.679998 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6427286d-a1ca-43ca-8ecb-bbdda45e360e-kube-api-access-zqq9t" (OuterVolumeSpecName: "kube-api-access-zqq9t") pod "6427286d-a1ca-43ca-8ecb-bbdda45e360e" (UID: "6427286d-a1ca-43ca-8ecb-bbdda45e360e"). InnerVolumeSpecName "kube-api-access-zqq9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.693221 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edb76b11-9129-495d-b44d-e998c3e8dceb-scripts" (OuterVolumeSpecName: "scripts") pod "edb76b11-9129-495d-b44d-e998c3e8dceb" (UID: "edb76b11-9129-495d-b44d-e998c3e8dceb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.707922 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edb76b11-9129-495d-b44d-e998c3e8dceb-kube-api-access-79x2b" (OuterVolumeSpecName: "kube-api-access-79x2b") pod "edb76b11-9129-495d-b44d-e998c3e8dceb" (UID: "edb76b11-9129-495d-b44d-e998c3e8dceb"). InnerVolumeSpecName "kube-api-access-79x2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.709727 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6427286d-a1ca-43ca-8ecb-bbdda45e360e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "6427286d-a1ca-43ca-8ecb-bbdda45e360e" (UID: "6427286d-a1ca-43ca-8ecb-bbdda45e360e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.769954 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edb76b11-9129-495d-b44d-e998c3e8dceb-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.769984 4934 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edb76b11-9129-495d-b44d-e998c3e8dceb-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.770022 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqq9t\" (UniqueName: \"kubernetes.io/projected/6427286d-a1ca-43ca-8ecb-bbdda45e360e-kube-api-access-zqq9t\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.770036 4934 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6427286d-a1ca-43ca-8ecb-bbdda45e360e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.770047 4934 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/edb76b11-9129-495d-b44d-e998c3e8dceb-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.770058 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79x2b\" (UniqueName: \"kubernetes.io/projected/edb76b11-9129-495d-b44d-e998c3e8dceb-kube-api-access-79x2b\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.789018 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.937850 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edb76b11-9129-495d-b44d-e998c3e8dceb-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "edb76b11-9129-495d-b44d-e998c3e8dceb" (UID: "edb76b11-9129-495d-b44d-e998c3e8dceb"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.945479 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57c45ff45d-ghdvr" Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.972400 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbxrb\" (UniqueName: \"kubernetes.io/projected/f2059861-459c-448a-933d-bc0d7660a9dc-kube-api-access-fbxrb\") pod \"f2059861-459c-448a-933d-bc0d7660a9dc\" (UID: \"f2059861-459c-448a-933d-bc0d7660a9dc\") " Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.972690 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-config\") pod \"f2059861-459c-448a-933d-bc0d7660a9dc\" (UID: \"f2059861-459c-448a-933d-bc0d7660a9dc\") " Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.972786 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-dns-swift-storage-0\") pod \"f2059861-459c-448a-933d-bc0d7660a9dc\" (UID: \"f2059861-459c-448a-933d-bc0d7660a9dc\") " Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.972928 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-ovsdbserver-sb\") pod \"f2059861-459c-448a-933d-bc0d7660a9dc\" (UID: \"f2059861-459c-448a-933d-bc0d7660a9dc\") " Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.972953 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-ovsdbserver-nb\") pod \"f2059861-459c-448a-933d-bc0d7660a9dc\" (UID: \"f2059861-459c-448a-933d-bc0d7660a9dc\") " Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.973094 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-dns-svc\") pod \"f2059861-459c-448a-933d-bc0d7660a9dc\" (UID: \"f2059861-459c-448a-933d-bc0d7660a9dc\") " Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.975040 4934 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/edb76b11-9129-495d-b44d-e998c3e8dceb-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:07 crc kubenswrapper[4934]: I1002 10:08:07.994688 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2059861-459c-448a-933d-bc0d7660a9dc-kube-api-access-fbxrb" (OuterVolumeSpecName: "kube-api-access-fbxrb") pod "f2059861-459c-448a-933d-bc0d7660a9dc" (UID: "f2059861-459c-448a-933d-bc0d7660a9dc"). InnerVolumeSpecName "kube-api-access-fbxrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.049484 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.077945 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckbgc\" (UniqueName: \"kubernetes.io/projected/a8799f51-9cec-454b-bdbe-860ce455baca-kube-api-access-ckbgc\") pod \"a8799f51-9cec-454b-bdbe-860ce455baca\" (UID: \"a8799f51-9cec-454b-bdbe-860ce455baca\") " Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.078000 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8799f51-9cec-454b-bdbe-860ce455baca-config-data\") pod \"a8799f51-9cec-454b-bdbe-860ce455baca\" (UID: \"a8799f51-9cec-454b-bdbe-860ce455baca\") " Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.078023 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a8799f51-9cec-454b-bdbe-860ce455baca-config-data-custom\") pod \"a8799f51-9cec-454b-bdbe-860ce455baca\" (UID: \"a8799f51-9cec-454b-bdbe-860ce455baca\") " Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.078055 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8799f51-9cec-454b-bdbe-860ce455baca-combined-ca-bundle\") pod \"a8799f51-9cec-454b-bdbe-860ce455baca\" (UID: \"a8799f51-9cec-454b-bdbe-860ce455baca\") " Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.078109 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8799f51-9cec-454b-bdbe-860ce455baca-logs\") pod \"a8799f51-9cec-454b-bdbe-860ce455baca\" (UID: \"a8799f51-9cec-454b-bdbe-860ce455baca\") " Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.080925 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8799f51-9cec-454b-bdbe-860ce455baca-logs" (OuterVolumeSpecName: "logs") pod "a8799f51-9cec-454b-bdbe-860ce455baca" (UID: "a8799f51-9cec-454b-bdbe-860ce455baca"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.101778 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbxrb\" (UniqueName: \"kubernetes.io/projected/f2059861-459c-448a-933d-bc0d7660a9dc-kube-api-access-fbxrb\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.107287 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8799f51-9cec-454b-bdbe-860ce455baca-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.126035 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8799f51-9cec-454b-bdbe-860ce455baca-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a8799f51-9cec-454b-bdbe-860ce455baca" (UID: "a8799f51-9cec-454b-bdbe-860ce455baca"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.128905 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8799f51-9cec-454b-bdbe-860ce455baca-kube-api-access-ckbgc" (OuterVolumeSpecName: "kube-api-access-ckbgc") pod "a8799f51-9cec-454b-bdbe-860ce455baca" (UID: "a8799f51-9cec-454b-bdbe-860ce455baca"). InnerVolumeSpecName "kube-api-access-ckbgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.135185 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6427286d-a1ca-43ca-8ecb-bbdda45e360e-config-data" (OuterVolumeSpecName: "config-data") pod "6427286d-a1ca-43ca-8ecb-bbdda45e360e" (UID: "6427286d-a1ca-43ca-8ecb-bbdda45e360e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.179763 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6427286d-a1ca-43ca-8ecb-bbdda45e360e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6427286d-a1ca-43ca-8ecb-bbdda45e360e" (UID: "6427286d-a1ca-43ca-8ecb-bbdda45e360e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.209811 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0d0eed2-217a-4dbc-906a-797c09b32081-config-data\") pod \"e0d0eed2-217a-4dbc-906a-797c09b32081\" (UID: \"e0d0eed2-217a-4dbc-906a-797c09b32081\") " Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.210195 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0d0eed2-217a-4dbc-906a-797c09b32081-config-data-custom\") pod \"e0d0eed2-217a-4dbc-906a-797c09b32081\" (UID: \"e0d0eed2-217a-4dbc-906a-797c09b32081\") " Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.210273 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0d0eed2-217a-4dbc-906a-797c09b32081-combined-ca-bundle\") pod \"e0d0eed2-217a-4dbc-906a-797c09b32081\" (UID: \"e0d0eed2-217a-4dbc-906a-797c09b32081\") " Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.210417 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mt2t\" (UniqueName: \"kubernetes.io/projected/e0d0eed2-217a-4dbc-906a-797c09b32081-kube-api-access-8mt2t\") pod \"e0d0eed2-217a-4dbc-906a-797c09b32081\" (UID: \"e0d0eed2-217a-4dbc-906a-797c09b32081\") " Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.210531 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0d0eed2-217a-4dbc-906a-797c09b32081-scripts\") pod \"e0d0eed2-217a-4dbc-906a-797c09b32081\" (UID: \"e0d0eed2-217a-4dbc-906a-797c09b32081\") " Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.210671 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0d0eed2-217a-4dbc-906a-797c09b32081-etc-machine-id\") pod \"e0d0eed2-217a-4dbc-906a-797c09b32081\" (UID: \"e0d0eed2-217a-4dbc-906a-797c09b32081\") " Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.211170 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckbgc\" (UniqueName: \"kubernetes.io/projected/a8799f51-9cec-454b-bdbe-860ce455baca-kube-api-access-ckbgc\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.211272 4934 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a8799f51-9cec-454b-bdbe-860ce455baca-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.211350 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6427286d-a1ca-43ca-8ecb-bbdda45e360e-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.211411 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6427286d-a1ca-43ca-8ecb-bbdda45e360e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.211496 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e0d0eed2-217a-4dbc-906a-797c09b32081-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e0d0eed2-217a-4dbc-906a-797c09b32081" (UID: "e0d0eed2-217a-4dbc-906a-797c09b32081"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.218542 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edb76b11-9129-495d-b44d-e998c3e8dceb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "edb76b11-9129-495d-b44d-e998c3e8dceb" (UID: "edb76b11-9129-495d-b44d-e998c3e8dceb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.224600 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8799f51-9cec-454b-bdbe-860ce455baca-config-data" (OuterVolumeSpecName: "config-data") pod "a8799f51-9cec-454b-bdbe-860ce455baca" (UID: "a8799f51-9cec-454b-bdbe-860ce455baca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.226031 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0d0eed2-217a-4dbc-906a-797c09b32081-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e0d0eed2-217a-4dbc-906a-797c09b32081" (UID: "e0d0eed2-217a-4dbc-906a-797c09b32081"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.227992 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0d0eed2-217a-4dbc-906a-797c09b32081-kube-api-access-8mt2t" (OuterVolumeSpecName: "kube-api-access-8mt2t") pod "e0d0eed2-217a-4dbc-906a-797c09b32081" (UID: "e0d0eed2-217a-4dbc-906a-797c09b32081"). InnerVolumeSpecName "kube-api-access-8mt2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.228153 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0d0eed2-217a-4dbc-906a-797c09b32081-scripts" (OuterVolumeSpecName: "scripts") pod "e0d0eed2-217a-4dbc-906a-797c09b32081" (UID: "e0d0eed2-217a-4dbc-906a-797c09b32081"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.235687 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8799f51-9cec-454b-bdbe-860ce455baca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a8799f51-9cec-454b-bdbe-860ce455baca" (UID: "a8799f51-9cec-454b-bdbe-860ce455baca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.250568 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f2059861-459c-448a-933d-bc0d7660a9dc" (UID: "f2059861-459c-448a-933d-bc0d7660a9dc"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.254554 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f2059861-459c-448a-933d-bc0d7660a9dc" (UID: "f2059861-459c-448a-933d-bc0d7660a9dc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.255935 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f2059861-459c-448a-933d-bc0d7660a9dc" (UID: "f2059861-459c-448a-933d-bc0d7660a9dc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.256183 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edb76b11-9129-495d-b44d-e998c3e8dceb-config-data" (OuterVolumeSpecName: "config-data") pod "edb76b11-9129-495d-b44d-e998c3e8dceb" (UID: "edb76b11-9129-495d-b44d-e998c3e8dceb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.278778 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f2059861-459c-448a-933d-bc0d7660a9dc" (UID: "f2059861-459c-448a-933d-bc0d7660a9dc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.289560 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-config" (OuterVolumeSpecName: "config") pod "f2059861-459c-448a-933d-bc0d7660a9dc" (UID: "f2059861-459c-448a-933d-bc0d7660a9dc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.296732 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0d0eed2-217a-4dbc-906a-797c09b32081-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e0d0eed2-217a-4dbc-906a-797c09b32081" (UID: "e0d0eed2-217a-4dbc-906a-797c09b32081"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.313516 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.313550 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.313561 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edb76b11-9129-495d-b44d-e998c3e8dceb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.313608 4934 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0d0eed2-217a-4dbc-906a-797c09b32081-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.313622 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.313632 4934 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0d0eed2-217a-4dbc-906a-797c09b32081-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.313643 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0d0eed2-217a-4dbc-906a-797c09b32081-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.313653 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8799f51-9cec-454b-bdbe-860ce455baca-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.313663 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8799f51-9cec-454b-bdbe-860ce455baca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.313672 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.313681 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edb76b11-9129-495d-b44d-e998c3e8dceb-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.313690 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mt2t\" (UniqueName: \"kubernetes.io/projected/e0d0eed2-217a-4dbc-906a-797c09b32081-kube-api-access-8mt2t\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.313701 4934 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f2059861-459c-448a-933d-bc0d7660a9dc-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.313709 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0d0eed2-217a-4dbc-906a-797c09b32081-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.334892 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0d0eed2-217a-4dbc-906a-797c09b32081-config-data" (OuterVolumeSpecName: "config-data") pod "e0d0eed2-217a-4dbc-906a-797c09b32081" (UID: "e0d0eed2-217a-4dbc-906a-797c09b32081"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.416077 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0d0eed2-217a-4dbc-906a-797c09b32081-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.439609 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.439674 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.439749 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.440560 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7b2842dc9e6894e9abab64461b922632ce5e3be7a7134bbbd340310d8f4b1d25"} pod="openshift-machine-config-operator/machine-config-daemon-djh5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.440646 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" containerID="cri-o://7b2842dc9e6894e9abab64461b922632ce5e3be7a7134bbbd340310d8f4b1d25" gracePeriod=600 Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.626789 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-bfc5bc4b9-crmsr" event={"ID":"a30daf4c-06f2-48f5-b57f-f6704fae13a9","Type":"ContainerStarted","Data":"65dbab7a03568a76645932c0b29b78b90ef3f3460bb8573a3ba9cd75761a9154"} Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.626835 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-bfc5bc4b9-crmsr" event={"ID":"a30daf4c-06f2-48f5-b57f-f6704fae13a9","Type":"ContainerStarted","Data":"ba2dbd7671783785d215106bb6b057dd861995948702cd2be58811e1a9b4c38e"} Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.627934 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.627961 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.631598 4934 generic.go:334] "Generic (PLEG): container finished" podID="71db06ef-05b0-4f58-b251-b27117a8500a" containerID="7b2842dc9e6894e9abab64461b922632ce5e3be7a7134bbbd340310d8f4b1d25" exitCode=0 Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.631675 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerDied","Data":"7b2842dc9e6894e9abab64461b922632ce5e3be7a7134bbbd340310d8f4b1d25"} Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.631710 4934 scope.go:117] "RemoveContainer" containerID="5567160fec03506e3dac2e60d3a06fa8b4a72b51448a03102f357a2e1f66bb4b" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.635935 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"9c99e264-05b3-4cd7-91ae-cabb6d10466a","Type":"ContainerStarted","Data":"7e1c034adb12673de1bc4203a0717c7690b9b6165dd60bea908b057b397ba31e"} Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.640233 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57c45ff45d-ghdvr" event={"ID":"a8799f51-9cec-454b-bdbe-860ce455baca","Type":"ContainerDied","Data":"425c56e98c096fa60ea560c5ecd4529df3356c6d41f8bf7052a06a014e538115"} Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.640476 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57c45ff45d-ghdvr" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.643961 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e0d0eed2-217a-4dbc-906a-797c09b32081","Type":"ContainerDied","Data":"b5bec96098ee1d10a714cbf2c1e87475985099d9a3c2f3f34b61fdccced4a950"} Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.644055 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.651803 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-bfc5bc4b9-crmsr" podStartSLOduration=8.651784235000001 podStartE2EDuration="8.651784235s" podCreationTimestamp="2025-10-02 10:08:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:08:08.65023363 +0000 UTC m=+1160.402875152" watchObservedRunningTime="2025-10-02 10:08:08.651784235 +0000 UTC m=+1160.404425757" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.656029 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" event={"ID":"f2059861-459c-448a-933d-bc0d7660a9dc","Type":"ContainerDied","Data":"a9b7c19121b2e90dd5c9edf16a5d9a6bae463406ba7ded086e41589b2d221c73"} Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.656143 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-685744b6b9-5mdh2" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.656906 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.664690 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jwpkw" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.684608 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.199672688 podStartE2EDuration="15.684570008s" podCreationTimestamp="2025-10-02 10:07:53 +0000 UTC" firstStartedPulling="2025-10-02 10:07:54.081348413 +0000 UTC m=+1145.833989925" lastFinishedPulling="2025-10-02 10:08:07.566245723 +0000 UTC m=+1159.318887245" observedRunningTime="2025-10-02 10:08:08.674762759 +0000 UTC m=+1160.427404281" watchObservedRunningTime="2025-10-02 10:08:08.684570008 +0000 UTC m=+1160.437211530" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.734445 4934 scope.go:117] "RemoveContainer" containerID="59ff67fae05185b501e4f0db278a33b889fc417ce780a16ef043c8e691f5fcbd" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.800952 4934 scope.go:117] "RemoveContainer" containerID="ee4233b97847dce506924805dd7b618d62016dea5d61be04baec7613cd33b597" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.827661 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-685744b6b9-5mdh2"] Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.851068 4934 scope.go:117] "RemoveContainer" containerID="8c229984e28e11867bf51914dc16c055d7e38ce173f5c33e9bf9b496ae9e06d5" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.867956 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-685744b6b9-5mdh2"] Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.887692 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.887890 4934 scope.go:117] "RemoveContainer" containerID="1bface884ab14709445ccfca87f0780807ad68ad0f60be0fc9cc86ff635fa242" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.913414 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.932172 4934 scope.go:117] "RemoveContainer" containerID="a2aa6bf2a4073de25d3a5fb820f9b97b8128dc4b37142dd94163513f5caff7bc" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.944937 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0d0eed2-217a-4dbc-906a-797c09b32081" path="/var/lib/kubelet/pods/e0d0eed2-217a-4dbc-906a-797c09b32081/volumes" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.945714 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2059861-459c-448a-933d-bc0d7660a9dc" path="/var/lib/kubelet/pods/f2059861-459c-448a-933d-bc0d7660a9dc/volumes" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.946501 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 10:08:08 crc kubenswrapper[4934]: E1002 10:08:08.946887 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2059861-459c-448a-933d-bc0d7660a9dc" containerName="dnsmasq-dns" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.946901 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2059861-459c-448a-933d-bc0d7660a9dc" containerName="dnsmasq-dns" Oct 02 10:08:08 crc kubenswrapper[4934]: E1002 10:08:08.946912 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0d0eed2-217a-4dbc-906a-797c09b32081" containerName="cinder-scheduler" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.946918 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0d0eed2-217a-4dbc-906a-797c09b32081" containerName="cinder-scheduler" Oct 02 10:08:08 crc kubenswrapper[4934]: E1002 10:08:08.946927 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0d0eed2-217a-4dbc-906a-797c09b32081" containerName="probe" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.946933 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0d0eed2-217a-4dbc-906a-797c09b32081" containerName="probe" Oct 02 10:08:08 crc kubenswrapper[4934]: E1002 10:08:08.946950 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edb76b11-9129-495d-b44d-e998c3e8dceb" containerName="ceilometer-central-agent" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.946956 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="edb76b11-9129-495d-b44d-e998c3e8dceb" containerName="ceilometer-central-agent" Oct 02 10:08:08 crc kubenswrapper[4934]: E1002 10:08:08.946964 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edb76b11-9129-495d-b44d-e998c3e8dceb" containerName="ceilometer-notification-agent" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.946969 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="edb76b11-9129-495d-b44d-e998c3e8dceb" containerName="ceilometer-notification-agent" Oct 02 10:08:08 crc kubenswrapper[4934]: E1002 10:08:08.946978 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2059861-459c-448a-933d-bc0d7660a9dc" containerName="init" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.946984 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2059861-459c-448a-933d-bc0d7660a9dc" containerName="init" Oct 02 10:08:08 crc kubenswrapper[4934]: E1002 10:08:08.946991 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6427286d-a1ca-43ca-8ecb-bbdda45e360e" containerName="glance-db-sync" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.946996 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="6427286d-a1ca-43ca-8ecb-bbdda45e360e" containerName="glance-db-sync" Oct 02 10:08:08 crc kubenswrapper[4934]: E1002 10:08:08.947008 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8799f51-9cec-454b-bdbe-860ce455baca" containerName="barbican-api-log" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.947015 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8799f51-9cec-454b-bdbe-860ce455baca" containerName="barbican-api-log" Oct 02 10:08:08 crc kubenswrapper[4934]: E1002 10:08:08.947031 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edb76b11-9129-495d-b44d-e998c3e8dceb" containerName="sg-core" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.947039 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="edb76b11-9129-495d-b44d-e998c3e8dceb" containerName="sg-core" Oct 02 10:08:08 crc kubenswrapper[4934]: E1002 10:08:08.947046 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8799f51-9cec-454b-bdbe-860ce455baca" containerName="barbican-api" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.947052 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8799f51-9cec-454b-bdbe-860ce455baca" containerName="barbican-api" Oct 02 10:08:08 crc kubenswrapper[4934]: E1002 10:08:08.947072 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edb76b11-9129-495d-b44d-e998c3e8dceb" containerName="proxy-httpd" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.947080 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="edb76b11-9129-495d-b44d-e998c3e8dceb" containerName="proxy-httpd" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.947249 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="edb76b11-9129-495d-b44d-e998c3e8dceb" containerName="ceilometer-notification-agent" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.947263 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="edb76b11-9129-495d-b44d-e998c3e8dceb" containerName="ceilometer-central-agent" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.947271 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0d0eed2-217a-4dbc-906a-797c09b32081" containerName="probe" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.947278 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0d0eed2-217a-4dbc-906a-797c09b32081" containerName="cinder-scheduler" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.947295 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8799f51-9cec-454b-bdbe-860ce455baca" containerName="barbican-api" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.947302 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="6427286d-a1ca-43ca-8ecb-bbdda45e360e" containerName="glance-db-sync" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.947311 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2059861-459c-448a-933d-bc0d7660a9dc" containerName="dnsmasq-dns" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.947317 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8799f51-9cec-454b-bdbe-860ce455baca" containerName="barbican-api-log" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.947331 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="edb76b11-9129-495d-b44d-e998c3e8dceb" containerName="sg-core" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.947338 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="edb76b11-9129-495d-b44d-e998c3e8dceb" containerName="proxy-httpd" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.952933 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.965375 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.975817 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.980854 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 02 10:08:08 crc kubenswrapper[4934]: I1002 10:08:08.996653 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.011000 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-57c45ff45d-ghdvr"] Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.011232 4934 scope.go:117] "RemoveContainer" containerID="58914f2a2ca9fc802b30dedd381ae8e1b8d1bc5e534d9e74a0b61b4ece557ab9" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.023133 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-57c45ff45d-ghdvr"] Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.030396 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.036496 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.042937 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.044422 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.066826 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.098586 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64c6bc45f9-ghf82"] Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.100016 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64c6bc45f9-ghf82" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.124308 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64c6bc45f9-ghf82"] Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.134481 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94a95a16-6bdd-4c08-8340-13fd85455b7a-scripts\") pod \"cinder-scheduler-0\" (UID: \"94a95a16-6bdd-4c08-8340-13fd85455b7a\") " pod="openstack/cinder-scheduler-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.134563 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c3a1512e-c161-4264-ba36-148c10f52a36-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c3a1512e-c161-4264-ba36-148c10f52a36\") " pod="openstack/ceilometer-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.134601 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3a1512e-c161-4264-ba36-148c10f52a36-config-data\") pod \"ceilometer-0\" (UID: \"c3a1512e-c161-4264-ba36-148c10f52a36\") " pod="openstack/ceilometer-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.134622 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zxvn\" (UniqueName: \"kubernetes.io/projected/c3a1512e-c161-4264-ba36-148c10f52a36-kube-api-access-2zxvn\") pod \"ceilometer-0\" (UID: \"c3a1512e-c161-4264-ba36-148c10f52a36\") " pod="openstack/ceilometer-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.134641 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3a1512e-c161-4264-ba36-148c10f52a36-run-httpd\") pod \"ceilometer-0\" (UID: \"c3a1512e-c161-4264-ba36-148c10f52a36\") " pod="openstack/ceilometer-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.134687 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3a1512e-c161-4264-ba36-148c10f52a36-scripts\") pod \"ceilometer-0\" (UID: \"c3a1512e-c161-4264-ba36-148c10f52a36\") " pod="openstack/ceilometer-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.134734 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3a1512e-c161-4264-ba36-148c10f52a36-log-httpd\") pod \"ceilometer-0\" (UID: \"c3a1512e-c161-4264-ba36-148c10f52a36\") " pod="openstack/ceilometer-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.134770 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/94a95a16-6bdd-4c08-8340-13fd85455b7a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"94a95a16-6bdd-4c08-8340-13fd85455b7a\") " pod="openstack/cinder-scheduler-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.134789 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n69mk\" (UniqueName: \"kubernetes.io/projected/94a95a16-6bdd-4c08-8340-13fd85455b7a-kube-api-access-n69mk\") pod \"cinder-scheduler-0\" (UID: \"94a95a16-6bdd-4c08-8340-13fd85455b7a\") " pod="openstack/cinder-scheduler-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.134812 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a95a16-6bdd-4c08-8340-13fd85455b7a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"94a95a16-6bdd-4c08-8340-13fd85455b7a\") " pod="openstack/cinder-scheduler-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.134826 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94a95a16-6bdd-4c08-8340-13fd85455b7a-config-data\") pod \"cinder-scheduler-0\" (UID: \"94a95a16-6bdd-4c08-8340-13fd85455b7a\") " pod="openstack/cinder-scheduler-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.134840 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3a1512e-c161-4264-ba36-148c10f52a36-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c3a1512e-c161-4264-ba36-148c10f52a36\") " pod="openstack/ceilometer-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.134869 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/94a95a16-6bdd-4c08-8340-13fd85455b7a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"94a95a16-6bdd-4c08-8340-13fd85455b7a\") " pod="openstack/cinder-scheduler-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.246715 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n69mk\" (UniqueName: \"kubernetes.io/projected/94a95a16-6bdd-4c08-8340-13fd85455b7a-kube-api-access-n69mk\") pod \"cinder-scheduler-0\" (UID: \"94a95a16-6bdd-4c08-8340-13fd85455b7a\") " pod="openstack/cinder-scheduler-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.247080 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/94a95a16-6bdd-4c08-8340-13fd85455b7a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"94a95a16-6bdd-4c08-8340-13fd85455b7a\") " pod="openstack/cinder-scheduler-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.247128 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-dns-swift-storage-0\") pod \"dnsmasq-dns-64c6bc45f9-ghf82\" (UID: \"7bd9ac9d-9468-45ae-996a-e2b742164c36\") " pod="openstack/dnsmasq-dns-64c6bc45f9-ghf82" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.247158 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a95a16-6bdd-4c08-8340-13fd85455b7a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"94a95a16-6bdd-4c08-8340-13fd85455b7a\") " pod="openstack/cinder-scheduler-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.247183 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94a95a16-6bdd-4c08-8340-13fd85455b7a-config-data\") pod \"cinder-scheduler-0\" (UID: \"94a95a16-6bdd-4c08-8340-13fd85455b7a\") " pod="openstack/cinder-scheduler-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.247204 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3a1512e-c161-4264-ba36-148c10f52a36-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c3a1512e-c161-4264-ba36-148c10f52a36\") " pod="openstack/ceilometer-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.247245 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/94a95a16-6bdd-4c08-8340-13fd85455b7a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"94a95a16-6bdd-4c08-8340-13fd85455b7a\") " pod="openstack/cinder-scheduler-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.247293 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfj4n\" (UniqueName: \"kubernetes.io/projected/7bd9ac9d-9468-45ae-996a-e2b742164c36-kube-api-access-pfj4n\") pod \"dnsmasq-dns-64c6bc45f9-ghf82\" (UID: \"7bd9ac9d-9468-45ae-996a-e2b742164c36\") " pod="openstack/dnsmasq-dns-64c6bc45f9-ghf82" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.247317 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94a95a16-6bdd-4c08-8340-13fd85455b7a-scripts\") pod \"cinder-scheduler-0\" (UID: \"94a95a16-6bdd-4c08-8340-13fd85455b7a\") " pod="openstack/cinder-scheduler-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.247357 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c3a1512e-c161-4264-ba36-148c10f52a36-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c3a1512e-c161-4264-ba36-148c10f52a36\") " pod="openstack/ceilometer-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.247378 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3a1512e-c161-4264-ba36-148c10f52a36-config-data\") pod \"ceilometer-0\" (UID: \"c3a1512e-c161-4264-ba36-148c10f52a36\") " pod="openstack/ceilometer-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.247401 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-ovsdbserver-nb\") pod \"dnsmasq-dns-64c6bc45f9-ghf82\" (UID: \"7bd9ac9d-9468-45ae-996a-e2b742164c36\") " pod="openstack/dnsmasq-dns-64c6bc45f9-ghf82" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.247425 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zxvn\" (UniqueName: \"kubernetes.io/projected/c3a1512e-c161-4264-ba36-148c10f52a36-kube-api-access-2zxvn\") pod \"ceilometer-0\" (UID: \"c3a1512e-c161-4264-ba36-148c10f52a36\") " pod="openstack/ceilometer-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.247452 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3a1512e-c161-4264-ba36-148c10f52a36-run-httpd\") pod \"ceilometer-0\" (UID: \"c3a1512e-c161-4264-ba36-148c10f52a36\") " pod="openstack/ceilometer-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.247493 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-dns-svc\") pod \"dnsmasq-dns-64c6bc45f9-ghf82\" (UID: \"7bd9ac9d-9468-45ae-996a-e2b742164c36\") " pod="openstack/dnsmasq-dns-64c6bc45f9-ghf82" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.247501 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/94a95a16-6bdd-4c08-8340-13fd85455b7a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"94a95a16-6bdd-4c08-8340-13fd85455b7a\") " pod="openstack/cinder-scheduler-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.247537 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3a1512e-c161-4264-ba36-148c10f52a36-scripts\") pod \"ceilometer-0\" (UID: \"c3a1512e-c161-4264-ba36-148c10f52a36\") " pod="openstack/ceilometer-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.247608 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-config\") pod \"dnsmasq-dns-64c6bc45f9-ghf82\" (UID: \"7bd9ac9d-9468-45ae-996a-e2b742164c36\") " pod="openstack/dnsmasq-dns-64c6bc45f9-ghf82" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.247647 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-ovsdbserver-sb\") pod \"dnsmasq-dns-64c6bc45f9-ghf82\" (UID: \"7bd9ac9d-9468-45ae-996a-e2b742164c36\") " pod="openstack/dnsmasq-dns-64c6bc45f9-ghf82" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.247693 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3a1512e-c161-4264-ba36-148c10f52a36-log-httpd\") pod \"ceilometer-0\" (UID: \"c3a1512e-c161-4264-ba36-148c10f52a36\") " pod="openstack/ceilometer-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.248140 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3a1512e-c161-4264-ba36-148c10f52a36-log-httpd\") pod \"ceilometer-0\" (UID: \"c3a1512e-c161-4264-ba36-148c10f52a36\") " pod="openstack/ceilometer-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.254249 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3a1512e-c161-4264-ba36-148c10f52a36-run-httpd\") pod \"ceilometer-0\" (UID: \"c3a1512e-c161-4264-ba36-148c10f52a36\") " pod="openstack/ceilometer-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.257037 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c3a1512e-c161-4264-ba36-148c10f52a36-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c3a1512e-c161-4264-ba36-148c10f52a36\") " pod="openstack/ceilometer-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.260207 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a95a16-6bdd-4c08-8340-13fd85455b7a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"94a95a16-6bdd-4c08-8340-13fd85455b7a\") " pod="openstack/cinder-scheduler-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.261140 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/94a95a16-6bdd-4c08-8340-13fd85455b7a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"94a95a16-6bdd-4c08-8340-13fd85455b7a\") " pod="openstack/cinder-scheduler-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.261978 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94a95a16-6bdd-4c08-8340-13fd85455b7a-scripts\") pod \"cinder-scheduler-0\" (UID: \"94a95a16-6bdd-4c08-8340-13fd85455b7a\") " pod="openstack/cinder-scheduler-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.263972 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-9z9t6" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.264079 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3a1512e-c161-4264-ba36-148c10f52a36-config-data\") pod \"ceilometer-0\" (UID: \"c3a1512e-c161-4264-ba36-148c10f52a36\") " pod="openstack/ceilometer-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.271286 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94a95a16-6bdd-4c08-8340-13fd85455b7a-config-data\") pod \"cinder-scheduler-0\" (UID: \"94a95a16-6bdd-4c08-8340-13fd85455b7a\") " pod="openstack/cinder-scheduler-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.273480 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3a1512e-c161-4264-ba36-148c10f52a36-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c3a1512e-c161-4264-ba36-148c10f52a36\") " pod="openstack/ceilometer-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.274310 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3a1512e-c161-4264-ba36-148c10f52a36-scripts\") pod \"ceilometer-0\" (UID: \"c3a1512e-c161-4264-ba36-148c10f52a36\") " pod="openstack/ceilometer-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.281649 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zxvn\" (UniqueName: \"kubernetes.io/projected/c3a1512e-c161-4264-ba36-148c10f52a36-kube-api-access-2zxvn\") pod \"ceilometer-0\" (UID: \"c3a1512e-c161-4264-ba36-148c10f52a36\") " pod="openstack/ceilometer-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.287649 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n69mk\" (UniqueName: \"kubernetes.io/projected/94a95a16-6bdd-4c08-8340-13fd85455b7a-kube-api-access-n69mk\") pod \"cinder-scheduler-0\" (UID: \"94a95a16-6bdd-4c08-8340-13fd85455b7a\") " pod="openstack/cinder-scheduler-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.315153 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.350161 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-dns-swift-storage-0\") pod \"dnsmasq-dns-64c6bc45f9-ghf82\" (UID: \"7bd9ac9d-9468-45ae-996a-e2b742164c36\") " pod="openstack/dnsmasq-dns-64c6bc45f9-ghf82" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.350292 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfj4n\" (UniqueName: \"kubernetes.io/projected/7bd9ac9d-9468-45ae-996a-e2b742164c36-kube-api-access-pfj4n\") pod \"dnsmasq-dns-64c6bc45f9-ghf82\" (UID: \"7bd9ac9d-9468-45ae-996a-e2b742164c36\") " pod="openstack/dnsmasq-dns-64c6bc45f9-ghf82" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.350369 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-ovsdbserver-nb\") pod \"dnsmasq-dns-64c6bc45f9-ghf82\" (UID: \"7bd9ac9d-9468-45ae-996a-e2b742164c36\") " pod="openstack/dnsmasq-dns-64c6bc45f9-ghf82" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.350441 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-dns-svc\") pod \"dnsmasq-dns-64c6bc45f9-ghf82\" (UID: \"7bd9ac9d-9468-45ae-996a-e2b742164c36\") " pod="openstack/dnsmasq-dns-64c6bc45f9-ghf82" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.350523 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-config\") pod \"dnsmasq-dns-64c6bc45f9-ghf82\" (UID: \"7bd9ac9d-9468-45ae-996a-e2b742164c36\") " pod="openstack/dnsmasq-dns-64c6bc45f9-ghf82" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.350597 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-ovsdbserver-sb\") pod \"dnsmasq-dns-64c6bc45f9-ghf82\" (UID: \"7bd9ac9d-9468-45ae-996a-e2b742164c36\") " pod="openstack/dnsmasq-dns-64c6bc45f9-ghf82" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.352118 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-ovsdbserver-sb\") pod \"dnsmasq-dns-64c6bc45f9-ghf82\" (UID: \"7bd9ac9d-9468-45ae-996a-e2b742164c36\") " pod="openstack/dnsmasq-dns-64c6bc45f9-ghf82" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.352498 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-dns-svc\") pod \"dnsmasq-dns-64c6bc45f9-ghf82\" (UID: \"7bd9ac9d-9468-45ae-996a-e2b742164c36\") " pod="openstack/dnsmasq-dns-64c6bc45f9-ghf82" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.352715 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-config\") pod \"dnsmasq-dns-64c6bc45f9-ghf82\" (UID: \"7bd9ac9d-9468-45ae-996a-e2b742164c36\") " pod="openstack/dnsmasq-dns-64c6bc45f9-ghf82" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.352898 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-dns-swift-storage-0\") pod \"dnsmasq-dns-64c6bc45f9-ghf82\" (UID: \"7bd9ac9d-9468-45ae-996a-e2b742164c36\") " pod="openstack/dnsmasq-dns-64c6bc45f9-ghf82" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.353362 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-ovsdbserver-nb\") pod \"dnsmasq-dns-64c6bc45f9-ghf82\" (UID: \"7bd9ac9d-9468-45ae-996a-e2b742164c36\") " pod="openstack/dnsmasq-dns-64c6bc45f9-ghf82" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.380513 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfj4n\" (UniqueName: \"kubernetes.io/projected/7bd9ac9d-9468-45ae-996a-e2b742164c36-kube-api-access-pfj4n\") pod \"dnsmasq-dns-64c6bc45f9-ghf82\" (UID: \"7bd9ac9d-9468-45ae-996a-e2b742164c36\") " pod="openstack/dnsmasq-dns-64c6bc45f9-ghf82" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.439122 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.452380 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c03e58b-5a31-4295-ba05-d219daff78b1-combined-ca-bundle\") pod \"3c03e58b-5a31-4295-ba05-d219daff78b1\" (UID: \"3c03e58b-5a31-4295-ba05-d219daff78b1\") " Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.452449 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3c03e58b-5a31-4295-ba05-d219daff78b1-config\") pod \"3c03e58b-5a31-4295-ba05-d219daff78b1\" (UID: \"3c03e58b-5a31-4295-ba05-d219daff78b1\") " Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.452482 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7ht5\" (UniqueName: \"kubernetes.io/projected/3c03e58b-5a31-4295-ba05-d219daff78b1-kube-api-access-f7ht5\") pod \"3c03e58b-5a31-4295-ba05-d219daff78b1\" (UID: \"3c03e58b-5a31-4295-ba05-d219daff78b1\") " Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.457765 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c03e58b-5a31-4295-ba05-d219daff78b1-kube-api-access-f7ht5" (OuterVolumeSpecName: "kube-api-access-f7ht5") pod "3c03e58b-5a31-4295-ba05-d219daff78b1" (UID: "3c03e58b-5a31-4295-ba05-d219daff78b1"). InnerVolumeSpecName "kube-api-access-f7ht5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.482218 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64c6bc45f9-ghf82" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.510996 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c03e58b-5a31-4295-ba05-d219daff78b1-config" (OuterVolumeSpecName: "config") pod "3c03e58b-5a31-4295-ba05-d219daff78b1" (UID: "3c03e58b-5a31-4295-ba05-d219daff78b1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.541829 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c03e58b-5a31-4295-ba05-d219daff78b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3c03e58b-5a31-4295-ba05-d219daff78b1" (UID: "3c03e58b-5a31-4295-ba05-d219daff78b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.555120 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c03e58b-5a31-4295-ba05-d219daff78b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.555149 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/3c03e58b-5a31-4295-ba05-d219daff78b1-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.555159 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7ht5\" (UniqueName: \"kubernetes.io/projected/3c03e58b-5a31-4295-ba05-d219daff78b1-kube-api-access-f7ht5\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.670021 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-9z9t6" event={"ID":"3c03e58b-5a31-4295-ba05-d219daff78b1","Type":"ContainerDied","Data":"76ab97f80fb3e8659dbd1a3a9d52730dc747105f159de483bc059e6fc1a45f2e"} Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.670695 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76ab97f80fb3e8659dbd1a3a9d52730dc747105f159de483bc059e6fc1a45f2e" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.670993 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-9z9t6" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.697704 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerStarted","Data":"577a0f8f7b9147143b952a1ec195b9368082248053d4e85cde27a2757e47ab8c"} Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.836603 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64c6bc45f9-ghf82"] Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.859462 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55d67bcc45-h5d2q"] Oct 02 10:08:09 crc kubenswrapper[4934]: E1002 10:08:09.860181 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c03e58b-5a31-4295-ba05-d219daff78b1" containerName="neutron-db-sync" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.860207 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c03e58b-5a31-4295-ba05-d219daff78b1" containerName="neutron-db-sync" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.860423 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c03e58b-5a31-4295-ba05-d219daff78b1" containerName="neutron-db-sync" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.861733 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.875514 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55d67bcc45-h5d2q"] Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.910597 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.936349 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-856f7d44fb-8lgx8"] Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.937821 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-856f7d44fb-8lgx8" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.942346 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.942655 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.942850 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.943899 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-v9qtp" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.953489 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-856f7d44fb-8lgx8"] Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.961675 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-config\") pod \"dnsmasq-dns-55d67bcc45-h5d2q\" (UID: \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\") " pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.961723 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-dns-svc\") pod \"dnsmasq-dns-55d67bcc45-h5d2q\" (UID: \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\") " pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.961752 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-ovsdbserver-sb\") pod \"dnsmasq-dns-55d67bcc45-h5d2q\" (UID: \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\") " pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.961793 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-dns-swift-storage-0\") pod \"dnsmasq-dns-55d67bcc45-h5d2q\" (UID: \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\") " pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.961867 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-ovsdbserver-nb\") pod \"dnsmasq-dns-55d67bcc45-h5d2q\" (UID: \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\") " pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" Oct 02 10:08:09 crc kubenswrapper[4934]: I1002 10:08:09.961907 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2bln\" (UniqueName: \"kubernetes.io/projected/73e6e086-8443-46e5-89f8-21bc9bde6b4f-kube-api-access-g2bln\") pod \"dnsmasq-dns-55d67bcc45-h5d2q\" (UID: \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\") " pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.063421 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2bln\" (UniqueName: \"kubernetes.io/projected/73e6e086-8443-46e5-89f8-21bc9bde6b4f-kube-api-access-g2bln\") pod \"dnsmasq-dns-55d67bcc45-h5d2q\" (UID: \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\") " pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.063479 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d8702c0-4b10-41e5-b170-64a356553760-ovndb-tls-certs\") pod \"neutron-856f7d44fb-8lgx8\" (UID: \"9d8702c0-4b10-41e5-b170-64a356553760\") " pod="openstack/neutron-856f7d44fb-8lgx8" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.063507 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d8702c0-4b10-41e5-b170-64a356553760-combined-ca-bundle\") pod \"neutron-856f7d44fb-8lgx8\" (UID: \"9d8702c0-4b10-41e5-b170-64a356553760\") " pod="openstack/neutron-856f7d44fb-8lgx8" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.063536 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-config\") pod \"dnsmasq-dns-55d67bcc45-h5d2q\" (UID: \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\") " pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.063559 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9d8702c0-4b10-41e5-b170-64a356553760-httpd-config\") pod \"neutron-856f7d44fb-8lgx8\" (UID: \"9d8702c0-4b10-41e5-b170-64a356553760\") " pod="openstack/neutron-856f7d44fb-8lgx8" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.063591 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-dns-svc\") pod \"dnsmasq-dns-55d67bcc45-h5d2q\" (UID: \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\") " pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.063621 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-ovsdbserver-sb\") pod \"dnsmasq-dns-55d67bcc45-h5d2q\" (UID: \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\") " pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.063659 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-dns-swift-storage-0\") pod \"dnsmasq-dns-55d67bcc45-h5d2q\" (UID: \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\") " pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.063679 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdbmg\" (UniqueName: \"kubernetes.io/projected/9d8702c0-4b10-41e5-b170-64a356553760-kube-api-access-cdbmg\") pod \"neutron-856f7d44fb-8lgx8\" (UID: \"9d8702c0-4b10-41e5-b170-64a356553760\") " pod="openstack/neutron-856f7d44fb-8lgx8" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.063760 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-ovsdbserver-nb\") pod \"dnsmasq-dns-55d67bcc45-h5d2q\" (UID: \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\") " pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.063778 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9d8702c0-4b10-41e5-b170-64a356553760-config\") pod \"neutron-856f7d44fb-8lgx8\" (UID: \"9d8702c0-4b10-41e5-b170-64a356553760\") " pod="openstack/neutron-856f7d44fb-8lgx8" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.064606 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-config\") pod \"dnsmasq-dns-55d67bcc45-h5d2q\" (UID: \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\") " pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.064745 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-ovsdbserver-sb\") pod \"dnsmasq-dns-55d67bcc45-h5d2q\" (UID: \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\") " pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.068109 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-dns-swift-storage-0\") pod \"dnsmasq-dns-55d67bcc45-h5d2q\" (UID: \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\") " pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.073145 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-ovsdbserver-nb\") pod \"dnsmasq-dns-55d67bcc45-h5d2q\" (UID: \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\") " pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.073239 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-dns-svc\") pod \"dnsmasq-dns-55d67bcc45-h5d2q\" (UID: \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\") " pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.079285 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.086269 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2bln\" (UniqueName: \"kubernetes.io/projected/73e6e086-8443-46e5-89f8-21bc9bde6b4f-kube-api-access-g2bln\") pod \"dnsmasq-dns-55d67bcc45-h5d2q\" (UID: \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\") " pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.098339 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64c6bc45f9-ghf82"] Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.147841 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.151843 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.157288 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.162127 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.162389 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.162691 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-6f9q2" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.166700 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdbmg\" (UniqueName: \"kubernetes.io/projected/9d8702c0-4b10-41e5-b170-64a356553760-kube-api-access-cdbmg\") pod \"neutron-856f7d44fb-8lgx8\" (UID: \"9d8702c0-4b10-41e5-b170-64a356553760\") " pod="openstack/neutron-856f7d44fb-8lgx8" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.166830 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9d8702c0-4b10-41e5-b170-64a356553760-config\") pod \"neutron-856f7d44fb-8lgx8\" (UID: \"9d8702c0-4b10-41e5-b170-64a356553760\") " pod="openstack/neutron-856f7d44fb-8lgx8" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.166888 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d8702c0-4b10-41e5-b170-64a356553760-ovndb-tls-certs\") pod \"neutron-856f7d44fb-8lgx8\" (UID: \"9d8702c0-4b10-41e5-b170-64a356553760\") " pod="openstack/neutron-856f7d44fb-8lgx8" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.166916 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d8702c0-4b10-41e5-b170-64a356553760-combined-ca-bundle\") pod \"neutron-856f7d44fb-8lgx8\" (UID: \"9d8702c0-4b10-41e5-b170-64a356553760\") " pod="openstack/neutron-856f7d44fb-8lgx8" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.166955 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9d8702c0-4b10-41e5-b170-64a356553760-httpd-config\") pod \"neutron-856f7d44fb-8lgx8\" (UID: \"9d8702c0-4b10-41e5-b170-64a356553760\") " pod="openstack/neutron-856f7d44fb-8lgx8" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.172490 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9d8702c0-4b10-41e5-b170-64a356553760-httpd-config\") pod \"neutron-856f7d44fb-8lgx8\" (UID: \"9d8702c0-4b10-41e5-b170-64a356553760\") " pod="openstack/neutron-856f7d44fb-8lgx8" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.180262 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/9d8702c0-4b10-41e5-b170-64a356553760-config\") pod \"neutron-856f7d44fb-8lgx8\" (UID: \"9d8702c0-4b10-41e5-b170-64a356553760\") " pod="openstack/neutron-856f7d44fb-8lgx8" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.195118 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.199032 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d8702c0-4b10-41e5-b170-64a356553760-combined-ca-bundle\") pod \"neutron-856f7d44fb-8lgx8\" (UID: \"9d8702c0-4b10-41e5-b170-64a356553760\") " pod="openstack/neutron-856f7d44fb-8lgx8" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.199138 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d8702c0-4b10-41e5-b170-64a356553760-ovndb-tls-certs\") pod \"neutron-856f7d44fb-8lgx8\" (UID: \"9d8702c0-4b10-41e5-b170-64a356553760\") " pod="openstack/neutron-856f7d44fb-8lgx8" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.205798 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdbmg\" (UniqueName: \"kubernetes.io/projected/9d8702c0-4b10-41e5-b170-64a356553760-kube-api-access-cdbmg\") pod \"neutron-856f7d44fb-8lgx8\" (UID: \"9d8702c0-4b10-41e5-b170-64a356553760\") " pod="openstack/neutron-856f7d44fb-8lgx8" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.268998 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c213066-759b-4c42-9ec4-eacd0d3fb7da-logs\") pod \"glance-default-external-api-0\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.269041 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.269075 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0c213066-759b-4c42-9ec4-eacd0d3fb7da-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.269100 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fcd4\" (UniqueName: \"kubernetes.io/projected/0c213066-759b-4c42-9ec4-eacd0d3fb7da-kube-api-access-2fcd4\") pod \"glance-default-external-api-0\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.269156 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c213066-759b-4c42-9ec4-eacd0d3fb7da-scripts\") pod \"glance-default-external-api-0\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.269173 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c213066-759b-4c42-9ec4-eacd0d3fb7da-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.269210 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c213066-759b-4c42-9ec4-eacd0d3fb7da-config-data\") pod \"glance-default-external-api-0\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.278128 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.279739 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.286401 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.295635 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-856f7d44fb-8lgx8" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.308040 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.372457 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c44dbd08-806a-4a8a-b161-948521672bf0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.372509 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c44dbd08-806a-4a8a-b161-948521672bf0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.372540 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q87xv\" (UniqueName: \"kubernetes.io/projected/c44dbd08-806a-4a8a-b161-948521672bf0-kube-api-access-q87xv\") pod \"glance-default-internal-api-0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.372564 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c44dbd08-806a-4a8a-b161-948521672bf0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.372618 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c213066-759b-4c42-9ec4-eacd0d3fb7da-logs\") pod \"glance-default-external-api-0\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.372704 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.372790 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.372828 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0c213066-759b-4c42-9ec4-eacd0d3fb7da-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.372863 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c44dbd08-806a-4a8a-b161-948521672bf0-logs\") pod \"glance-default-internal-api-0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.372897 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fcd4\" (UniqueName: \"kubernetes.io/projected/0c213066-759b-4c42-9ec4-eacd0d3fb7da-kube-api-access-2fcd4\") pod \"glance-default-external-api-0\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.372923 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c44dbd08-806a-4a8a-b161-948521672bf0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.372959 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c213066-759b-4c42-9ec4-eacd0d3fb7da-logs\") pod \"glance-default-external-api-0\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.373169 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0c213066-759b-4c42-9ec4-eacd0d3fb7da-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.373249 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.373600 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c213066-759b-4c42-9ec4-eacd0d3fb7da-scripts\") pod \"glance-default-external-api-0\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.373626 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c213066-759b-4c42-9ec4-eacd0d3fb7da-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.373713 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c213066-759b-4c42-9ec4-eacd0d3fb7da-config-data\") pod \"glance-default-external-api-0\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.389707 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c213066-759b-4c42-9ec4-eacd0d3fb7da-config-data\") pod \"glance-default-external-api-0\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.390103 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fcd4\" (UniqueName: \"kubernetes.io/projected/0c213066-759b-4c42-9ec4-eacd0d3fb7da-kube-api-access-2fcd4\") pod \"glance-default-external-api-0\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.390431 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c213066-759b-4c42-9ec4-eacd0d3fb7da-scripts\") pod \"glance-default-external-api-0\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.394345 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c213066-759b-4c42-9ec4-eacd0d3fb7da-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.420528 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.475622 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.475968 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c44dbd08-806a-4a8a-b161-948521672bf0-logs\") pod \"glance-default-internal-api-0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.476003 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c44dbd08-806a-4a8a-b161-948521672bf0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.476127 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c44dbd08-806a-4a8a-b161-948521672bf0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.476171 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c44dbd08-806a-4a8a-b161-948521672bf0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.476190 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q87xv\" (UniqueName: \"kubernetes.io/projected/c44dbd08-806a-4a8a-b161-948521672bf0-kube-api-access-q87xv\") pod \"glance-default-internal-api-0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.476215 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c44dbd08-806a-4a8a-b161-948521672bf0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.476999 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.479351 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c44dbd08-806a-4a8a-b161-948521672bf0-logs\") pod \"glance-default-internal-api-0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.479740 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c44dbd08-806a-4a8a-b161-948521672bf0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.492317 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c44dbd08-806a-4a8a-b161-948521672bf0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.493906 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c44dbd08-806a-4a8a-b161-948521672bf0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.500165 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.507699 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c44dbd08-806a-4a8a-b161-948521672bf0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.534341 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q87xv\" (UniqueName: \"kubernetes.io/projected/c44dbd08-806a-4a8a-b161-948521672bf0-kube-api-access-q87xv\") pod \"glance-default-internal-api-0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.568777 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.617610 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.762809 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3a1512e-c161-4264-ba36-148c10f52a36","Type":"ContainerStarted","Data":"553f83e6f2e0bb757bb167a63920f7ca1f53ee05888db0dd64a03851c737b6ed"} Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.790509 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64c6bc45f9-ghf82" event={"ID":"7bd9ac9d-9468-45ae-996a-e2b742164c36","Type":"ContainerStarted","Data":"b47a8904de46f69be77731772a25967a135f79ab739fb873a37e2ee2722a74a0"} Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.804944 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"94a95a16-6bdd-4c08-8340-13fd85455b7a","Type":"ContainerStarted","Data":"9998933aed61ac0b0458b31757054f9862222bb41ad966c21139956e6a6eb297"} Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.945456 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8799f51-9cec-454b-bdbe-860ce455baca" path="/var/lib/kubelet/pods/a8799f51-9cec-454b-bdbe-860ce455baca/volumes" Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.946247 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edb76b11-9129-495d-b44d-e998c3e8dceb" path="/var/lib/kubelet/pods/edb76b11-9129-495d-b44d-e998c3e8dceb/volumes" Oct 02 10:08:10 crc kubenswrapper[4934]: W1002 10:08:10.949787 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73e6e086_8443_46e5_89f8_21bc9bde6b4f.slice/crio-e868d7e6f05fe721e79b35bdd140cbdd9d0ae1b99d3357190404fab44cc1190e WatchSource:0}: Error finding container e868d7e6f05fe721e79b35bdd140cbdd9d0ae1b99d3357190404fab44cc1190e: Status 404 returned error can't find the container with id e868d7e6f05fe721e79b35bdd140cbdd9d0ae1b99d3357190404fab44cc1190e Oct 02 10:08:10 crc kubenswrapper[4934]: I1002 10:08:10.961101 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55d67bcc45-h5d2q"] Oct 02 10:08:11 crc kubenswrapper[4934]: I1002 10:08:11.024544 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-856f7d44fb-8lgx8"] Oct 02 10:08:11 crc kubenswrapper[4934]: I1002 10:08:11.149556 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 02 10:08:11 crc kubenswrapper[4934]: I1002 10:08:11.379885 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:08:11 crc kubenswrapper[4934]: I1002 10:08:11.431514 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:08:11 crc kubenswrapper[4934]: I1002 10:08:11.851144 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0c213066-759b-4c42-9ec4-eacd0d3fb7da","Type":"ContainerStarted","Data":"6b33e65b098bc3e0f7d69c31f10cd1b92a47b70880ee25560679b7dc1b60599d"} Oct 02 10:08:11 crc kubenswrapper[4934]: I1002 10:08:11.855525 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" event={"ID":"73e6e086-8443-46e5-89f8-21bc9bde6b4f","Type":"ContainerStarted","Data":"685b0edf8574c2f1879850eeb6fc75939e3235e6d1e40e86f77b4f04d51d5f5e"} Oct 02 10:08:11 crc kubenswrapper[4934]: I1002 10:08:11.855556 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" event={"ID":"73e6e086-8443-46e5-89f8-21bc9bde6b4f","Type":"ContainerStarted","Data":"e868d7e6f05fe721e79b35bdd140cbdd9d0ae1b99d3357190404fab44cc1190e"} Oct 02 10:08:11 crc kubenswrapper[4934]: I1002 10:08:11.865120 4934 generic.go:334] "Generic (PLEG): container finished" podID="7bd9ac9d-9468-45ae-996a-e2b742164c36" containerID="fb4e62f78585d3a5b3acc088fb32f93bab9c6fc76d51489ea80f70fc7bf2d246" exitCode=0 Oct 02 10:08:11 crc kubenswrapper[4934]: I1002 10:08:11.866866 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64c6bc45f9-ghf82" event={"ID":"7bd9ac9d-9468-45ae-996a-e2b742164c36","Type":"ContainerDied","Data":"fb4e62f78585d3a5b3acc088fb32f93bab9c6fc76d51489ea80f70fc7bf2d246"} Oct 02 10:08:11 crc kubenswrapper[4934]: I1002 10:08:11.892778 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-856f7d44fb-8lgx8" event={"ID":"9d8702c0-4b10-41e5-b170-64a356553760","Type":"ContainerStarted","Data":"2a58baac49d821442c171f05a2c3c9c9971843e59c5742bee94443c034a65a7b"} Oct 02 10:08:11 crc kubenswrapper[4934]: I1002 10:08:11.895989 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"94a95a16-6bdd-4c08-8340-13fd85455b7a","Type":"ContainerStarted","Data":"9c952faa06a3569c88d733912510a64a2634e55c3359c72da98dba68500b8acf"} Oct 02 10:08:11 crc kubenswrapper[4934]: I1002 10:08:11.917023 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3a1512e-c161-4264-ba36-148c10f52a36","Type":"ContainerStarted","Data":"41ac05160deb747befeacdc68c2f9f4a9b84267fe81e7a9f257a1cfbdd276ec2"} Oct 02 10:08:11 crc kubenswrapper[4934]: I1002 10:08:11.921379 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.612971 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64c6bc45f9-ghf82" Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.785628 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-dns-swift-storage-0\") pod \"7bd9ac9d-9468-45ae-996a-e2b742164c36\" (UID: \"7bd9ac9d-9468-45ae-996a-e2b742164c36\") " Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.785676 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-dns-svc\") pod \"7bd9ac9d-9468-45ae-996a-e2b742164c36\" (UID: \"7bd9ac9d-9468-45ae-996a-e2b742164c36\") " Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.785722 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-ovsdbserver-sb\") pod \"7bd9ac9d-9468-45ae-996a-e2b742164c36\" (UID: \"7bd9ac9d-9468-45ae-996a-e2b742164c36\") " Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.785808 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-config\") pod \"7bd9ac9d-9468-45ae-996a-e2b742164c36\" (UID: \"7bd9ac9d-9468-45ae-996a-e2b742164c36\") " Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.785838 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfj4n\" (UniqueName: \"kubernetes.io/projected/7bd9ac9d-9468-45ae-996a-e2b742164c36-kube-api-access-pfj4n\") pod \"7bd9ac9d-9468-45ae-996a-e2b742164c36\" (UID: \"7bd9ac9d-9468-45ae-996a-e2b742164c36\") " Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.785922 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-ovsdbserver-nb\") pod \"7bd9ac9d-9468-45ae-996a-e2b742164c36\" (UID: \"7bd9ac9d-9468-45ae-996a-e2b742164c36\") " Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.808841 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bd9ac9d-9468-45ae-996a-e2b742164c36-kube-api-access-pfj4n" (OuterVolumeSpecName: "kube-api-access-pfj4n") pod "7bd9ac9d-9468-45ae-996a-e2b742164c36" (UID: "7bd9ac9d-9468-45ae-996a-e2b742164c36"). InnerVolumeSpecName "kube-api-access-pfj4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.848097 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7bd9ac9d-9468-45ae-996a-e2b742164c36" (UID: "7bd9ac9d-9468-45ae-996a-e2b742164c36"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.852350 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7bd9ac9d-9468-45ae-996a-e2b742164c36" (UID: "7bd9ac9d-9468-45ae-996a-e2b742164c36"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.855064 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7bd9ac9d-9468-45ae-996a-e2b742164c36" (UID: "7bd9ac9d-9468-45ae-996a-e2b742164c36"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.871222 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7bd9ac9d-9468-45ae-996a-e2b742164c36" (UID: "7bd9ac9d-9468-45ae-996a-e2b742164c36"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.879504 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-config" (OuterVolumeSpecName: "config") pod "7bd9ac9d-9468-45ae-996a-e2b742164c36" (UID: "7bd9ac9d-9468-45ae-996a-e2b742164c36"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.888105 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.888138 4934 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.888149 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.888157 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.888168 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7bd9ac9d-9468-45ae-996a-e2b742164c36-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.888176 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfj4n\" (UniqueName: \"kubernetes.io/projected/7bd9ac9d-9468-45ae-996a-e2b742164c36-kube-api-access-pfj4n\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.948409 4934 generic.go:334] "Generic (PLEG): container finished" podID="73e6e086-8443-46e5-89f8-21bc9bde6b4f" containerID="685b0edf8574c2f1879850eeb6fc75939e3235e6d1e40e86f77b4f04d51d5f5e" exitCode=0 Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.954372 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" event={"ID":"73e6e086-8443-46e5-89f8-21bc9bde6b4f","Type":"ContainerDied","Data":"685b0edf8574c2f1879850eeb6fc75939e3235e6d1e40e86f77b4f04d51d5f5e"} Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.960568 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64c6bc45f9-ghf82" Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.960697 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64c6bc45f9-ghf82" event={"ID":"7bd9ac9d-9468-45ae-996a-e2b742164c36","Type":"ContainerDied","Data":"b47a8904de46f69be77731772a25967a135f79ab739fb873a37e2ee2722a74a0"} Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.960746 4934 scope.go:117] "RemoveContainer" containerID="fb4e62f78585d3a5b3acc088fb32f93bab9c6fc76d51489ea80f70fc7bf2d246" Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.975001 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-856f7d44fb-8lgx8" event={"ID":"9d8702c0-4b10-41e5-b170-64a356553760","Type":"ContainerStarted","Data":"a2462b4c5120af0b9bbf15be9bd9b839ac5c8d932d347d54f9c06988a93fd7dd"} Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.975054 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-856f7d44fb-8lgx8" event={"ID":"9d8702c0-4b10-41e5-b170-64a356553760","Type":"ContainerStarted","Data":"7b1b88de67d53c8cce24ae235bd6d265be9585b8b71ccf82999a91c8aecefc42"} Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.975890 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-856f7d44fb-8lgx8" Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.997256 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"94a95a16-6bdd-4c08-8340-13fd85455b7a","Type":"ContainerStarted","Data":"7f805c812aa2a019f6996c4840760eec127d8f206fe82dbfef04db5d8dadd02c"} Oct 02 10:08:12 crc kubenswrapper[4934]: I1002 10:08:12.999311 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-856f7d44fb-8lgx8" podStartSLOduration=3.9992904510000002 podStartE2EDuration="3.999290451s" podCreationTimestamp="2025-10-02 10:08:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:08:12.997094429 +0000 UTC m=+1164.749735951" watchObservedRunningTime="2025-10-02 10:08:12.999290451 +0000 UTC m=+1164.751931973" Oct 02 10:08:13 crc kubenswrapper[4934]: I1002 10:08:13.027885 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c44dbd08-806a-4a8a-b161-948521672bf0","Type":"ContainerStarted","Data":"2ee8224dba91e72a2f7e1a4eab43ecdc514cb4c8f82d81164e2f06b8d48f3406"} Oct 02 10:08:13 crc kubenswrapper[4934]: I1002 10:08:13.065513 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3a1512e-c161-4264-ba36-148c10f52a36","Type":"ContainerStarted","Data":"01490ed90f471b4ad5147dc9eb7945067b7b6559aad2a7f47a8e593590da3985"} Oct 02 10:08:13 crc kubenswrapper[4934]: I1002 10:08:13.076112 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64c6bc45f9-ghf82"] Oct 02 10:08:13 crc kubenswrapper[4934]: I1002 10:08:13.085900 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0c213066-759b-4c42-9ec4-eacd0d3fb7da","Type":"ContainerStarted","Data":"4564b124b3ab6cb3fa9cafd7d9d389812f895dc73896b36224405dbcf07488c8"} Oct 02 10:08:13 crc kubenswrapper[4934]: I1002 10:08:13.124357 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64c6bc45f9-ghf82"] Oct 02 10:08:13 crc kubenswrapper[4934]: I1002 10:08:13.126077 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.126058545 podStartE2EDuration="5.126058545s" podCreationTimestamp="2025-10-02 10:08:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:08:13.085716764 +0000 UTC m=+1164.838358286" watchObservedRunningTime="2025-10-02 10:08:13.126058545 +0000 UTC m=+1164.878700067" Oct 02 10:08:13 crc kubenswrapper[4934]: I1002 10:08:13.644472 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:08:14 crc kubenswrapper[4934]: I1002 10:08:14.102774 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c44dbd08-806a-4a8a-b161-948521672bf0","Type":"ContainerStarted","Data":"468b157b80a3c62eb661c23f720f5878fafd1d3f2c60efa7f7cfc70f79e86149"} Oct 02 10:08:14 crc kubenswrapper[4934]: I1002 10:08:14.103157 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c44dbd08-806a-4a8a-b161-948521672bf0","Type":"ContainerStarted","Data":"67523659df3fe960d227298c43246590e5f84ec3cedd20013e2c9d13985cb2d2"} Oct 02 10:08:14 crc kubenswrapper[4934]: I1002 10:08:14.103175 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="c44dbd08-806a-4a8a-b161-948521672bf0" containerName="glance-log" containerID="cri-o://67523659df3fe960d227298c43246590e5f84ec3cedd20013e2c9d13985cb2d2" gracePeriod=30 Oct 02 10:08:14 crc kubenswrapper[4934]: I1002 10:08:14.103279 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="c44dbd08-806a-4a8a-b161-948521672bf0" containerName="glance-httpd" containerID="cri-o://468b157b80a3c62eb661c23f720f5878fafd1d3f2c60efa7f7cfc70f79e86149" gracePeriod=30 Oct 02 10:08:14 crc kubenswrapper[4934]: I1002 10:08:14.114818 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3a1512e-c161-4264-ba36-148c10f52a36","Type":"ContainerStarted","Data":"51def09dc4784bd05fa87ea6c929cfb1876b26f52ab5b324580b3d692ee0eb54"} Oct 02 10:08:14 crc kubenswrapper[4934]: I1002 10:08:14.133637 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0c213066-759b-4c42-9ec4-eacd0d3fb7da","Type":"ContainerStarted","Data":"d12ee260e2491ac8a0403671f630f6145e072ccf03895c49538c37ab8d7495e1"} Oct 02 10:08:14 crc kubenswrapper[4934]: I1002 10:08:14.133857 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="0c213066-759b-4c42-9ec4-eacd0d3fb7da" containerName="glance-log" containerID="cri-o://4564b124b3ab6cb3fa9cafd7d9d389812f895dc73896b36224405dbcf07488c8" gracePeriod=30 Oct 02 10:08:14 crc kubenswrapper[4934]: I1002 10:08:14.134344 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="0c213066-759b-4c42-9ec4-eacd0d3fb7da" containerName="glance-httpd" containerID="cri-o://d12ee260e2491ac8a0403671f630f6145e072ccf03895c49538c37ab8d7495e1" gracePeriod=30 Oct 02 10:08:14 crc kubenswrapper[4934]: I1002 10:08:14.167824 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.167796668 podStartE2EDuration="5.167796668s" podCreationTimestamp="2025-10-02 10:08:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:08:14.166188361 +0000 UTC m=+1165.918829883" watchObservedRunningTime="2025-10-02 10:08:14.167796668 +0000 UTC m=+1165.920438200" Oct 02 10:08:14 crc kubenswrapper[4934]: I1002 10:08:14.171289 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" event={"ID":"73e6e086-8443-46e5-89f8-21bc9bde6b4f","Type":"ContainerStarted","Data":"867e46f0e9e66fc2f4cda4cf5adfa64cd1fa5ad8df0d05ee0a276ebbba2ce317"} Oct 02 10:08:14 crc kubenswrapper[4934]: I1002 10:08:14.171716 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" Oct 02 10:08:14 crc kubenswrapper[4934]: I1002 10:08:14.172083 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.172060159 podStartE2EDuration="5.172060159s" podCreationTimestamp="2025-10-02 10:08:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:08:14.133694615 +0000 UTC m=+1165.886336137" watchObservedRunningTime="2025-10-02 10:08:14.172060159 +0000 UTC m=+1165.924701691" Oct 02 10:08:14 crc kubenswrapper[4934]: I1002 10:08:14.228407 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" podStartSLOduration=5.228377134 podStartE2EDuration="5.228377134s" podCreationTimestamp="2025-10-02 10:08:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:08:14.216116915 +0000 UTC m=+1165.968758437" watchObservedRunningTime="2025-10-02 10:08:14.228377134 +0000 UTC m=+1165.981018646" Oct 02 10:08:14 crc kubenswrapper[4934]: I1002 10:08:14.316952 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 02 10:08:14 crc kubenswrapper[4934]: I1002 10:08:14.903903 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 10:08:14 crc kubenswrapper[4934]: I1002 10:08:14.935657 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bd9ac9d-9468-45ae-996a-e2b742164c36" path="/var/lib/kubelet/pods/7bd9ac9d-9468-45ae-996a-e2b742164c36/volumes" Oct 02 10:08:14 crc kubenswrapper[4934]: I1002 10:08:14.962827 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.052317 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c213066-759b-4c42-9ec4-eacd0d3fb7da-combined-ca-bundle\") pod \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") " Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.052370 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c213066-759b-4c42-9ec4-eacd0d3fb7da-logs\") pod \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") " Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.052396 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c44dbd08-806a-4a8a-b161-948521672bf0-scripts\") pod \"c44dbd08-806a-4a8a-b161-948521672bf0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") " Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.052447 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q87xv\" (UniqueName: \"kubernetes.io/projected/c44dbd08-806a-4a8a-b161-948521672bf0-kube-api-access-q87xv\") pod \"c44dbd08-806a-4a8a-b161-948521672bf0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") " Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.052485 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c213066-759b-4c42-9ec4-eacd0d3fb7da-config-data\") pod \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") " Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.052521 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c44dbd08-806a-4a8a-b161-948521672bf0-combined-ca-bundle\") pod \"c44dbd08-806a-4a8a-b161-948521672bf0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") " Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.052617 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0c213066-759b-4c42-9ec4-eacd0d3fb7da-httpd-run\") pod \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") " Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.052650 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fcd4\" (UniqueName: \"kubernetes.io/projected/0c213066-759b-4c42-9ec4-eacd0d3fb7da-kube-api-access-2fcd4\") pod \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") " Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.052700 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c44dbd08-806a-4a8a-b161-948521672bf0-httpd-run\") pod \"c44dbd08-806a-4a8a-b161-948521672bf0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") " Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.052787 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c44dbd08-806a-4a8a-b161-948521672bf0-config-data\") pod \"c44dbd08-806a-4a8a-b161-948521672bf0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") " Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.052812 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c213066-759b-4c42-9ec4-eacd0d3fb7da-scripts\") pod \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") " Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.052833 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c44dbd08-806a-4a8a-b161-948521672bf0-logs\") pod \"c44dbd08-806a-4a8a-b161-948521672bf0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") " Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.052857 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\" (UID: \"0c213066-759b-4c42-9ec4-eacd0d3fb7da\") " Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.052899 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"c44dbd08-806a-4a8a-b161-948521672bf0\" (UID: \"c44dbd08-806a-4a8a-b161-948521672bf0\") " Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.054210 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c213066-759b-4c42-9ec4-eacd0d3fb7da-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0c213066-759b-4c42-9ec4-eacd0d3fb7da" (UID: "0c213066-759b-4c42-9ec4-eacd0d3fb7da"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.059496 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "c44dbd08-806a-4a8a-b161-948521672bf0" (UID: "c44dbd08-806a-4a8a-b161-948521672bf0"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.069002 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c44dbd08-806a-4a8a-b161-948521672bf0-logs" (OuterVolumeSpecName: "logs") pod "c44dbd08-806a-4a8a-b161-948521672bf0" (UID: "c44dbd08-806a-4a8a-b161-948521672bf0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.069351 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c213066-759b-4c42-9ec4-eacd0d3fb7da-logs" (OuterVolumeSpecName: "logs") pod "0c213066-759b-4c42-9ec4-eacd0d3fb7da" (UID: "0c213066-759b-4c42-9ec4-eacd0d3fb7da"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.070279 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c44dbd08-806a-4a8a-b161-948521672bf0-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c44dbd08-806a-4a8a-b161-948521672bf0" (UID: "c44dbd08-806a-4a8a-b161-948521672bf0"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.087146 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c213066-759b-4c42-9ec4-eacd0d3fb7da-scripts" (OuterVolumeSpecName: "scripts") pod "0c213066-759b-4c42-9ec4-eacd0d3fb7da" (UID: "0c213066-759b-4c42-9ec4-eacd0d3fb7da"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.087248 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c44dbd08-806a-4a8a-b161-948521672bf0-scripts" (OuterVolumeSpecName: "scripts") pod "c44dbd08-806a-4a8a-b161-948521672bf0" (UID: "c44dbd08-806a-4a8a-b161-948521672bf0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.094533 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c213066-759b-4c42-9ec4-eacd0d3fb7da-kube-api-access-2fcd4" (OuterVolumeSpecName: "kube-api-access-2fcd4") pod "0c213066-759b-4c42-9ec4-eacd0d3fb7da" (UID: "0c213066-759b-4c42-9ec4-eacd0d3fb7da"). InnerVolumeSpecName "kube-api-access-2fcd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.101853 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "0c213066-759b-4c42-9ec4-eacd0d3fb7da" (UID: "0c213066-759b-4c42-9ec4-eacd0d3fb7da"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.102022 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c44dbd08-806a-4a8a-b161-948521672bf0-kube-api-access-q87xv" (OuterVolumeSpecName: "kube-api-access-q87xv") pod "c44dbd08-806a-4a8a-b161-948521672bf0" (UID: "c44dbd08-806a-4a8a-b161-948521672bf0"). InnerVolumeSpecName "kube-api-access-q87xv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.137704 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c44dbd08-806a-4a8a-b161-948521672bf0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c44dbd08-806a-4a8a-b161-948521672bf0" (UID: "c44dbd08-806a-4a8a-b161-948521672bf0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.158048 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c213066-759b-4c42-9ec4-eacd0d3fb7da-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.158093 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c44dbd08-806a-4a8a-b161-948521672bf0-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.158126 4934 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.158156 4934 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.158169 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c213066-759b-4c42-9ec4-eacd0d3fb7da-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.158181 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c44dbd08-806a-4a8a-b161-948521672bf0-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.158193 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q87xv\" (UniqueName: \"kubernetes.io/projected/c44dbd08-806a-4a8a-b161-948521672bf0-kube-api-access-q87xv\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.158206 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c44dbd08-806a-4a8a-b161-948521672bf0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.158219 4934 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0c213066-759b-4c42-9ec4-eacd0d3fb7da-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.158229 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fcd4\" (UniqueName: \"kubernetes.io/projected/0c213066-759b-4c42-9ec4-eacd0d3fb7da-kube-api-access-2fcd4\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.158240 4934 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c44dbd08-806a-4a8a-b161-948521672bf0-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.160127 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c213066-759b-4c42-9ec4-eacd0d3fb7da-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c213066-759b-4c42-9ec4-eacd0d3fb7da" (UID: "0c213066-759b-4c42-9ec4-eacd0d3fb7da"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.166712 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c213066-759b-4c42-9ec4-eacd0d3fb7da-config-data" (OuterVolumeSpecName: "config-data") pod "0c213066-759b-4c42-9ec4-eacd0d3fb7da" (UID: "0c213066-759b-4c42-9ec4-eacd0d3fb7da"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.183845 4934 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.194684 4934 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.196958 4934 generic.go:334] "Generic (PLEG): container finished" podID="c44dbd08-806a-4a8a-b161-948521672bf0" containerID="468b157b80a3c62eb661c23f720f5878fafd1d3f2c60efa7f7cfc70f79e86149" exitCode=143 Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.196987 4934 generic.go:334] "Generic (PLEG): container finished" podID="c44dbd08-806a-4a8a-b161-948521672bf0" containerID="67523659df3fe960d227298c43246590e5f84ec3cedd20013e2c9d13985cb2d2" exitCode=143 Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.197039 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c44dbd08-806a-4a8a-b161-948521672bf0","Type":"ContainerDied","Data":"468b157b80a3c62eb661c23f720f5878fafd1d3f2c60efa7f7cfc70f79e86149"} Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.197075 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c44dbd08-806a-4a8a-b161-948521672bf0","Type":"ContainerDied","Data":"67523659df3fe960d227298c43246590e5f84ec3cedd20013e2c9d13985cb2d2"} Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.197087 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c44dbd08-806a-4a8a-b161-948521672bf0","Type":"ContainerDied","Data":"2ee8224dba91e72a2f7e1a4eab43ecdc514cb4c8f82d81164e2f06b8d48f3406"} Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.197084 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.197102 4934 scope.go:117] "RemoveContainer" containerID="468b157b80a3c62eb661c23f720f5878fafd1d3f2c60efa7f7cfc70f79e86149" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.199791 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c44dbd08-806a-4a8a-b161-948521672bf0-config-data" (OuterVolumeSpecName: "config-data") pod "c44dbd08-806a-4a8a-b161-948521672bf0" (UID: "c44dbd08-806a-4a8a-b161-948521672bf0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.210263 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3a1512e-c161-4264-ba36-148c10f52a36","Type":"ContainerStarted","Data":"353a5f848560344ee82aa27e6d2845a5d78d06029b45a9ff5d536eda488efade"} Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.211514 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.214186 4934 generic.go:334] "Generic (PLEG): container finished" podID="0c213066-759b-4c42-9ec4-eacd0d3fb7da" containerID="d12ee260e2491ac8a0403671f630f6145e072ccf03895c49538c37ab8d7495e1" exitCode=143 Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.214228 4934 generic.go:334] "Generic (PLEG): container finished" podID="0c213066-759b-4c42-9ec4-eacd0d3fb7da" containerID="4564b124b3ab6cb3fa9cafd7d9d389812f895dc73896b36224405dbcf07488c8" exitCode=143 Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.217248 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.219611 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0c213066-759b-4c42-9ec4-eacd0d3fb7da","Type":"ContainerDied","Data":"d12ee260e2491ac8a0403671f630f6145e072ccf03895c49538c37ab8d7495e1"} Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.219646 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0c213066-759b-4c42-9ec4-eacd0d3fb7da","Type":"ContainerDied","Data":"4564b124b3ab6cb3fa9cafd7d9d389812f895dc73896b36224405dbcf07488c8"} Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.219657 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0c213066-759b-4c42-9ec4-eacd0d3fb7da","Type":"ContainerDied","Data":"6b33e65b098bc3e0f7d69c31f10cd1b92a47b70880ee25560679b7dc1b60599d"} Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.236377 4934 scope.go:117] "RemoveContainer" containerID="67523659df3fe960d227298c43246590e5f84ec3cedd20013e2c9d13985cb2d2" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.239179 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.73808089 podStartE2EDuration="7.239164104s" podCreationTimestamp="2025-10-02 10:08:08 +0000 UTC" firstStartedPulling="2025-10-02 10:08:10.097768399 +0000 UTC m=+1161.850409921" lastFinishedPulling="2025-10-02 10:08:14.598851623 +0000 UTC m=+1166.351493135" observedRunningTime="2025-10-02 10:08:15.234180022 +0000 UTC m=+1166.986821534" watchObservedRunningTime="2025-10-02 10:08:15.239164104 +0000 UTC m=+1166.991805626" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.267842 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.269511 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c44dbd08-806a-4a8a-b161-948521672bf0-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.269534 4934 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.269542 4934 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.269549 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c213066-759b-4c42-9ec4-eacd0d3fb7da-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.269558 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c213066-759b-4c42-9ec4-eacd0d3fb7da-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.274790 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.312370 4934 scope.go:117] "RemoveContainer" containerID="468b157b80a3c62eb661c23f720f5878fafd1d3f2c60efa7f7cfc70f79e86149" Oct 02 10:08:15 crc kubenswrapper[4934]: E1002 10:08:15.314071 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"468b157b80a3c62eb661c23f720f5878fafd1d3f2c60efa7f7cfc70f79e86149\": container with ID starting with 468b157b80a3c62eb661c23f720f5878fafd1d3f2c60efa7f7cfc70f79e86149 not found: ID does not exist" containerID="468b157b80a3c62eb661c23f720f5878fafd1d3f2c60efa7f7cfc70f79e86149" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.314109 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"468b157b80a3c62eb661c23f720f5878fafd1d3f2c60efa7f7cfc70f79e86149"} err="failed to get container status \"468b157b80a3c62eb661c23f720f5878fafd1d3f2c60efa7f7cfc70f79e86149\": rpc error: code = NotFound desc = could not find container \"468b157b80a3c62eb661c23f720f5878fafd1d3f2c60efa7f7cfc70f79e86149\": container with ID starting with 468b157b80a3c62eb661c23f720f5878fafd1d3f2c60efa7f7cfc70f79e86149 not found: ID does not exist" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.314135 4934 scope.go:117] "RemoveContainer" containerID="67523659df3fe960d227298c43246590e5f84ec3cedd20013e2c9d13985cb2d2" Oct 02 10:08:15 crc kubenswrapper[4934]: E1002 10:08:15.314336 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67523659df3fe960d227298c43246590e5f84ec3cedd20013e2c9d13985cb2d2\": container with ID starting with 67523659df3fe960d227298c43246590e5f84ec3cedd20013e2c9d13985cb2d2 not found: ID does not exist" containerID="67523659df3fe960d227298c43246590e5f84ec3cedd20013e2c9d13985cb2d2" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.314355 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67523659df3fe960d227298c43246590e5f84ec3cedd20013e2c9d13985cb2d2"} err="failed to get container status \"67523659df3fe960d227298c43246590e5f84ec3cedd20013e2c9d13985cb2d2\": rpc error: code = NotFound desc = could not find container \"67523659df3fe960d227298c43246590e5f84ec3cedd20013e2c9d13985cb2d2\": container with ID starting with 67523659df3fe960d227298c43246590e5f84ec3cedd20013e2c9d13985cb2d2 not found: ID does not exist" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.314370 4934 scope.go:117] "RemoveContainer" containerID="468b157b80a3c62eb661c23f720f5878fafd1d3f2c60efa7f7cfc70f79e86149" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.314522 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"468b157b80a3c62eb661c23f720f5878fafd1d3f2c60efa7f7cfc70f79e86149"} err="failed to get container status \"468b157b80a3c62eb661c23f720f5878fafd1d3f2c60efa7f7cfc70f79e86149\": rpc error: code = NotFound desc = could not find container \"468b157b80a3c62eb661c23f720f5878fafd1d3f2c60efa7f7cfc70f79e86149\": container with ID starting with 468b157b80a3c62eb661c23f720f5878fafd1d3f2c60efa7f7cfc70f79e86149 not found: ID does not exist" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.314540 4934 scope.go:117] "RemoveContainer" containerID="67523659df3fe960d227298c43246590e5f84ec3cedd20013e2c9d13985cb2d2" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.314710 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67523659df3fe960d227298c43246590e5f84ec3cedd20013e2c9d13985cb2d2"} err="failed to get container status \"67523659df3fe960d227298c43246590e5f84ec3cedd20013e2c9d13985cb2d2\": rpc error: code = NotFound desc = could not find container \"67523659df3fe960d227298c43246590e5f84ec3cedd20013e2c9d13985cb2d2\": container with ID starting with 67523659df3fe960d227298c43246590e5f84ec3cedd20013e2c9d13985cb2d2 not found: ID does not exist" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.314727 4934 scope.go:117] "RemoveContainer" containerID="d12ee260e2491ac8a0403671f630f6145e072ccf03895c49538c37ab8d7495e1" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.329833 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:08:15 crc kubenswrapper[4934]: E1002 10:08:15.330218 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c44dbd08-806a-4a8a-b161-948521672bf0" containerName="glance-log" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.330228 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c44dbd08-806a-4a8a-b161-948521672bf0" containerName="glance-log" Oct 02 10:08:15 crc kubenswrapper[4934]: E1002 10:08:15.330247 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c213066-759b-4c42-9ec4-eacd0d3fb7da" containerName="glance-log" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.330253 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c213066-759b-4c42-9ec4-eacd0d3fb7da" containerName="glance-log" Oct 02 10:08:15 crc kubenswrapper[4934]: E1002 10:08:15.330269 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c44dbd08-806a-4a8a-b161-948521672bf0" containerName="glance-httpd" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.330289 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c44dbd08-806a-4a8a-b161-948521672bf0" containerName="glance-httpd" Oct 02 10:08:15 crc kubenswrapper[4934]: E1002 10:08:15.330302 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c213066-759b-4c42-9ec4-eacd0d3fb7da" containerName="glance-httpd" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.330308 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c213066-759b-4c42-9ec4-eacd0d3fb7da" containerName="glance-httpd" Oct 02 10:08:15 crc kubenswrapper[4934]: E1002 10:08:15.330330 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bd9ac9d-9468-45ae-996a-e2b742164c36" containerName="init" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.330336 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bd9ac9d-9468-45ae-996a-e2b742164c36" containerName="init" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.330508 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="c44dbd08-806a-4a8a-b161-948521672bf0" containerName="glance-log" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.330523 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bd9ac9d-9468-45ae-996a-e2b742164c36" containerName="init" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.330539 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c213066-759b-4c42-9ec4-eacd0d3fb7da" containerName="glance-httpd" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.330546 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="c44dbd08-806a-4a8a-b161-948521672bf0" containerName="glance-httpd" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.330558 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c213066-759b-4c42-9ec4-eacd0d3fb7da" containerName="glance-log" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.331410 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.334238 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.345051 4934 scope.go:117] "RemoveContainer" containerID="4564b124b3ab6cb3fa9cafd7d9d389812f895dc73896b36224405dbcf07488c8" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.363640 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.378773 4934 scope.go:117] "RemoveContainer" containerID="d12ee260e2491ac8a0403671f630f6145e072ccf03895c49538c37ab8d7495e1" Oct 02 10:08:15 crc kubenswrapper[4934]: E1002 10:08:15.380160 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d12ee260e2491ac8a0403671f630f6145e072ccf03895c49538c37ab8d7495e1\": container with ID starting with d12ee260e2491ac8a0403671f630f6145e072ccf03895c49538c37ab8d7495e1 not found: ID does not exist" containerID="d12ee260e2491ac8a0403671f630f6145e072ccf03895c49538c37ab8d7495e1" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.380189 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d12ee260e2491ac8a0403671f630f6145e072ccf03895c49538c37ab8d7495e1"} err="failed to get container status \"d12ee260e2491ac8a0403671f630f6145e072ccf03895c49538c37ab8d7495e1\": rpc error: code = NotFound desc = could not find container \"d12ee260e2491ac8a0403671f630f6145e072ccf03895c49538c37ab8d7495e1\": container with ID starting with d12ee260e2491ac8a0403671f630f6145e072ccf03895c49538c37ab8d7495e1 not found: ID does not exist" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.380212 4934 scope.go:117] "RemoveContainer" containerID="4564b124b3ab6cb3fa9cafd7d9d389812f895dc73896b36224405dbcf07488c8" Oct 02 10:08:15 crc kubenswrapper[4934]: E1002 10:08:15.384965 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4564b124b3ab6cb3fa9cafd7d9d389812f895dc73896b36224405dbcf07488c8\": container with ID starting with 4564b124b3ab6cb3fa9cafd7d9d389812f895dc73896b36224405dbcf07488c8 not found: ID does not exist" containerID="4564b124b3ab6cb3fa9cafd7d9d389812f895dc73896b36224405dbcf07488c8" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.385005 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4564b124b3ab6cb3fa9cafd7d9d389812f895dc73896b36224405dbcf07488c8"} err="failed to get container status \"4564b124b3ab6cb3fa9cafd7d9d389812f895dc73896b36224405dbcf07488c8\": rpc error: code = NotFound desc = could not find container \"4564b124b3ab6cb3fa9cafd7d9d389812f895dc73896b36224405dbcf07488c8\": container with ID starting with 4564b124b3ab6cb3fa9cafd7d9d389812f895dc73896b36224405dbcf07488c8 not found: ID does not exist" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.385031 4934 scope.go:117] "RemoveContainer" containerID="d12ee260e2491ac8a0403671f630f6145e072ccf03895c49538c37ab8d7495e1" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.387848 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d12ee260e2491ac8a0403671f630f6145e072ccf03895c49538c37ab8d7495e1"} err="failed to get container status \"d12ee260e2491ac8a0403671f630f6145e072ccf03895c49538c37ab8d7495e1\": rpc error: code = NotFound desc = could not find container \"d12ee260e2491ac8a0403671f630f6145e072ccf03895c49538c37ab8d7495e1\": container with ID starting with d12ee260e2491ac8a0403671f630f6145e072ccf03895c49538c37ab8d7495e1 not found: ID does not exist" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.387880 4934 scope.go:117] "RemoveContainer" containerID="4564b124b3ab6cb3fa9cafd7d9d389812f895dc73896b36224405dbcf07488c8" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.388277 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4564b124b3ab6cb3fa9cafd7d9d389812f895dc73896b36224405dbcf07488c8"} err="failed to get container status \"4564b124b3ab6cb3fa9cafd7d9d389812f895dc73896b36224405dbcf07488c8\": rpc error: code = NotFound desc = could not find container \"4564b124b3ab6cb3fa9cafd7d9d389812f895dc73896b36224405dbcf07488c8\": container with ID starting with 4564b124b3ab6cb3fa9cafd7d9d389812f895dc73896b36224405dbcf07488c8 not found: ID does not exist" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.474839 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cf37f71-cb02-4053-9e82-71a041cf8463-logs\") pod \"glance-default-external-api-0\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.474887 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sn4l\" (UniqueName: \"kubernetes.io/projected/7cf37f71-cb02-4053-9e82-71a041cf8463-kube-api-access-9sn4l\") pod \"glance-default-external-api-0\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.474920 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.474971 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cf37f71-cb02-4053-9e82-71a041cf8463-config-data\") pod \"glance-default-external-api-0\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.475019 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7cf37f71-cb02-4053-9e82-71a041cf8463-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.475042 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cf37f71-cb02-4053-9e82-71a041cf8463-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.475075 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cf37f71-cb02-4053-9e82-71a041cf8463-scripts\") pod \"glance-default-external-api-0\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.513961 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.514989 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.536892 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.570161 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.579761 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cf37f71-cb02-4053-9e82-71a041cf8463-logs\") pod \"glance-default-external-api-0\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.579825 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sn4l\" (UniqueName: \"kubernetes.io/projected/7cf37f71-cb02-4053-9e82-71a041cf8463-kube-api-access-9sn4l\") pod \"glance-default-external-api-0\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.579876 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.579956 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cf37f71-cb02-4053-9e82-71a041cf8463-config-data\") pod \"glance-default-external-api-0\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.580038 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7cf37f71-cb02-4053-9e82-71a041cf8463-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.580068 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cf37f71-cb02-4053-9e82-71a041cf8463-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.580130 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cf37f71-cb02-4053-9e82-71a041cf8463-scripts\") pod \"glance-default-external-api-0\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.582926 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.583124 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7cf37f71-cb02-4053-9e82-71a041cf8463-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.584372 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.587803 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cf37f71-cb02-4053-9e82-71a041cf8463-scripts\") pod \"glance-default-external-api-0\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.588650 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cf37f71-cb02-4053-9e82-71a041cf8463-config-data\") pod \"glance-default-external-api-0\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.587134 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cf37f71-cb02-4053-9e82-71a041cf8463-logs\") pod \"glance-default-external-api-0\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.610372 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cf37f71-cb02-4053-9e82-71a041cf8463-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.617304 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.626401 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sn4l\" (UniqueName: \"kubernetes.io/projected/7cf37f71-cb02-4053-9e82-71a041cf8463-kube-api-access-9sn4l\") pod \"glance-default-external-api-0\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.653774 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.669388 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.724767 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.727643 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.731752 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.767292 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.784597 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e434dd5-17db-4f21-87d5-e41d09717e88-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.784650 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e434dd5-17db-4f21-87d5-e41d09717e88-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.784753 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3e434dd5-17db-4f21-87d5-e41d09717e88-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.784862 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jmsc\" (UniqueName: \"kubernetes.io/projected/3e434dd5-17db-4f21-87d5-e41d09717e88-kube-api-access-2jmsc\") pod \"glance-default-internal-api-0\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.784945 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e434dd5-17db-4f21-87d5-e41d09717e88-logs\") pod \"glance-default-internal-api-0\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.785013 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e434dd5-17db-4f21-87d5-e41d09717e88-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.785059 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.888682 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e434dd5-17db-4f21-87d5-e41d09717e88-logs\") pod \"glance-default-internal-api-0\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.888780 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e434dd5-17db-4f21-87d5-e41d09717e88-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.888837 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.889457 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e434dd5-17db-4f21-87d5-e41d09717e88-logs\") pod \"glance-default-internal-api-0\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.889568 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e434dd5-17db-4f21-87d5-e41d09717e88-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.889635 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e434dd5-17db-4f21-87d5-e41d09717e88-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.889677 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3e434dd5-17db-4f21-87d5-e41d09717e88-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.889730 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jmsc\" (UniqueName: \"kubernetes.io/projected/3e434dd5-17db-4f21-87d5-e41d09717e88-kube-api-access-2jmsc\") pod \"glance-default-internal-api-0\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.899266 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.910273 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e434dd5-17db-4f21-87d5-e41d09717e88-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.914007 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e434dd5-17db-4f21-87d5-e41d09717e88-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.921627 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3e434dd5-17db-4f21-87d5-e41d09717e88-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.951983 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e434dd5-17db-4f21-87d5-e41d09717e88-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.954933 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:08:15 crc kubenswrapper[4934]: E1002 10:08:15.956027 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[glance kube-api-access-2jmsc], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/glance-default-internal-api-0" podUID="3e434dd5-17db-4f21-87d5-e41d09717e88" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.972597 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jmsc\" (UniqueName: \"kubernetes.io/projected/3e434dd5-17db-4f21-87d5-e41d09717e88-kube-api-access-2jmsc\") pod \"glance-default-internal-api-0\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:15 crc kubenswrapper[4934]: I1002 10:08:15.978554 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.033087 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.088627 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-57456ffd97-rqc8b"] Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.090099 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-57456ffd97-rqc8b" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.094046 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.099937 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-57456ffd97-rqc8b"] Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.100448 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.235401 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.245984 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95sg5\" (UniqueName: \"kubernetes.io/projected/6d79c694-2253-419e-b63b-4a38884ac57e-kube-api-access-95sg5\") pod \"neutron-57456ffd97-rqc8b\" (UID: \"6d79c694-2253-419e-b63b-4a38884ac57e\") " pod="openstack/neutron-57456ffd97-rqc8b" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.246085 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-config\") pod \"neutron-57456ffd97-rqc8b\" (UID: \"6d79c694-2253-419e-b63b-4a38884ac57e\") " pod="openstack/neutron-57456ffd97-rqc8b" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.246141 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-internal-tls-certs\") pod \"neutron-57456ffd97-rqc8b\" (UID: \"6d79c694-2253-419e-b63b-4a38884ac57e\") " pod="openstack/neutron-57456ffd97-rqc8b" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.246289 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-public-tls-certs\") pod \"neutron-57456ffd97-rqc8b\" (UID: \"6d79c694-2253-419e-b63b-4a38884ac57e\") " pod="openstack/neutron-57456ffd97-rqc8b" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.246369 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-ovndb-tls-certs\") pod \"neutron-57456ffd97-rqc8b\" (UID: \"6d79c694-2253-419e-b63b-4a38884ac57e\") " pod="openstack/neutron-57456ffd97-rqc8b" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.246454 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-combined-ca-bundle\") pod \"neutron-57456ffd97-rqc8b\" (UID: \"6d79c694-2253-419e-b63b-4a38884ac57e\") " pod="openstack/neutron-57456ffd97-rqc8b" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.246518 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-httpd-config\") pod \"neutron-57456ffd97-rqc8b\" (UID: \"6d79c694-2253-419e-b63b-4a38884ac57e\") " pod="openstack/neutron-57456ffd97-rqc8b" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.250611 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.348249 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e434dd5-17db-4f21-87d5-e41d09717e88-logs\") pod \"3e434dd5-17db-4f21-87d5-e41d09717e88\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") " Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.348310 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3e434dd5-17db-4f21-87d5-e41d09717e88-httpd-run\") pod \"3e434dd5-17db-4f21-87d5-e41d09717e88\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") " Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.348773 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e434dd5-17db-4f21-87d5-e41d09717e88-logs" (OuterVolumeSpecName: "logs") pod "3e434dd5-17db-4f21-87d5-e41d09717e88" (UID: "3e434dd5-17db-4f21-87d5-e41d09717e88"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.348847 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e434dd5-17db-4f21-87d5-e41d09717e88-scripts\") pod \"3e434dd5-17db-4f21-87d5-e41d09717e88\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") " Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.348905 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"3e434dd5-17db-4f21-87d5-e41d09717e88\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") " Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.348940 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e434dd5-17db-4f21-87d5-e41d09717e88-combined-ca-bundle\") pod \"3e434dd5-17db-4f21-87d5-e41d09717e88\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") " Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.348975 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jmsc\" (UniqueName: \"kubernetes.io/projected/3e434dd5-17db-4f21-87d5-e41d09717e88-kube-api-access-2jmsc\") pod \"3e434dd5-17db-4f21-87d5-e41d09717e88\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") " Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.349053 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e434dd5-17db-4f21-87d5-e41d09717e88-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3e434dd5-17db-4f21-87d5-e41d09717e88" (UID: "3e434dd5-17db-4f21-87d5-e41d09717e88"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.349062 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e434dd5-17db-4f21-87d5-e41d09717e88-config-data\") pod \"3e434dd5-17db-4f21-87d5-e41d09717e88\" (UID: \"3e434dd5-17db-4f21-87d5-e41d09717e88\") " Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.349339 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-ovndb-tls-certs\") pod \"neutron-57456ffd97-rqc8b\" (UID: \"6d79c694-2253-419e-b63b-4a38884ac57e\") " pod="openstack/neutron-57456ffd97-rqc8b" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.349846 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-combined-ca-bundle\") pod \"neutron-57456ffd97-rqc8b\" (UID: \"6d79c694-2253-419e-b63b-4a38884ac57e\") " pod="openstack/neutron-57456ffd97-rqc8b" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.349904 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-httpd-config\") pod \"neutron-57456ffd97-rqc8b\" (UID: \"6d79c694-2253-419e-b63b-4a38884ac57e\") " pod="openstack/neutron-57456ffd97-rqc8b" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.350104 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95sg5\" (UniqueName: \"kubernetes.io/projected/6d79c694-2253-419e-b63b-4a38884ac57e-kube-api-access-95sg5\") pod \"neutron-57456ffd97-rqc8b\" (UID: \"6d79c694-2253-419e-b63b-4a38884ac57e\") " pod="openstack/neutron-57456ffd97-rqc8b" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.350131 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-config\") pod \"neutron-57456ffd97-rqc8b\" (UID: \"6d79c694-2253-419e-b63b-4a38884ac57e\") " pod="openstack/neutron-57456ffd97-rqc8b" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.350148 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-internal-tls-certs\") pod \"neutron-57456ffd97-rqc8b\" (UID: \"6d79c694-2253-419e-b63b-4a38884ac57e\") " pod="openstack/neutron-57456ffd97-rqc8b" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.350256 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-public-tls-certs\") pod \"neutron-57456ffd97-rqc8b\" (UID: \"6d79c694-2253-419e-b63b-4a38884ac57e\") " pod="openstack/neutron-57456ffd97-rqc8b" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.350322 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e434dd5-17db-4f21-87d5-e41d09717e88-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.350333 4934 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3e434dd5-17db-4f21-87d5-e41d09717e88-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.356318 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "3e434dd5-17db-4f21-87d5-e41d09717e88" (UID: "3e434dd5-17db-4f21-87d5-e41d09717e88"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.358947 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-public-tls-certs\") pod \"neutron-57456ffd97-rqc8b\" (UID: \"6d79c694-2253-419e-b63b-4a38884ac57e\") " pod="openstack/neutron-57456ffd97-rqc8b" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.359799 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e434dd5-17db-4f21-87d5-e41d09717e88-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3e434dd5-17db-4f21-87d5-e41d09717e88" (UID: "3e434dd5-17db-4f21-87d5-e41d09717e88"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.359883 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e434dd5-17db-4f21-87d5-e41d09717e88-config-data" (OuterVolumeSpecName: "config-data") pod "3e434dd5-17db-4f21-87d5-e41d09717e88" (UID: "3e434dd5-17db-4f21-87d5-e41d09717e88"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.360288 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-combined-ca-bundle\") pod \"neutron-57456ffd97-rqc8b\" (UID: \"6d79c694-2253-419e-b63b-4a38884ac57e\") " pod="openstack/neutron-57456ffd97-rqc8b" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.360682 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e434dd5-17db-4f21-87d5-e41d09717e88-kube-api-access-2jmsc" (OuterVolumeSpecName: "kube-api-access-2jmsc") pod "3e434dd5-17db-4f21-87d5-e41d09717e88" (UID: "3e434dd5-17db-4f21-87d5-e41d09717e88"). InnerVolumeSpecName "kube-api-access-2jmsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.361375 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e434dd5-17db-4f21-87d5-e41d09717e88-scripts" (OuterVolumeSpecName: "scripts") pod "3e434dd5-17db-4f21-87d5-e41d09717e88" (UID: "3e434dd5-17db-4f21-87d5-e41d09717e88"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.362236 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-config\") pod \"neutron-57456ffd97-rqc8b\" (UID: \"6d79c694-2253-419e-b63b-4a38884ac57e\") " pod="openstack/neutron-57456ffd97-rqc8b" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.364506 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-ovndb-tls-certs\") pod \"neutron-57456ffd97-rqc8b\" (UID: \"6d79c694-2253-419e-b63b-4a38884ac57e\") " pod="openstack/neutron-57456ffd97-rqc8b" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.364623 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-httpd-config\") pod \"neutron-57456ffd97-rqc8b\" (UID: \"6d79c694-2253-419e-b63b-4a38884ac57e\") " pod="openstack/neutron-57456ffd97-rqc8b" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.372360 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-internal-tls-certs\") pod \"neutron-57456ffd97-rqc8b\" (UID: \"6d79c694-2253-419e-b63b-4a38884ac57e\") " pod="openstack/neutron-57456ffd97-rqc8b" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.373365 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95sg5\" (UniqueName: \"kubernetes.io/projected/6d79c694-2253-419e-b63b-4a38884ac57e-kube-api-access-95sg5\") pod \"neutron-57456ffd97-rqc8b\" (UID: \"6d79c694-2253-419e-b63b-4a38884ac57e\") " pod="openstack/neutron-57456ffd97-rqc8b" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.388411 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.420175 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-57456ffd97-rqc8b" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.452561 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e434dd5-17db-4f21-87d5-e41d09717e88-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.452631 4934 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.452648 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e434dd5-17db-4f21-87d5-e41d09717e88-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.452681 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jmsc\" (UniqueName: \"kubernetes.io/projected/3e434dd5-17db-4f21-87d5-e41d09717e88-kube-api-access-2jmsc\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.452691 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e434dd5-17db-4f21-87d5-e41d09717e88-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.473521 4934 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.554844 4934 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.951072 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c213066-759b-4c42-9ec4-eacd0d3fb7da" path="/var/lib/kubelet/pods/0c213066-759b-4c42-9ec4-eacd0d3fb7da/volumes" Oct 02 10:08:16 crc kubenswrapper[4934]: I1002 10:08:16.957796 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c44dbd08-806a-4a8a-b161-948521672bf0" path="/var/lib/kubelet/pods/c44dbd08-806a-4a8a-b161-948521672bf0/volumes" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.075006 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-57456ffd97-rqc8b"] Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.248892 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-57456ffd97-rqc8b" event={"ID":"6d79c694-2253-419e-b63b-4a38884ac57e","Type":"ContainerStarted","Data":"868a863f516906649063d12e1987b02363fef789aab5c137eb1a8f1869fc16a2"} Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.250725 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.251028 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7cf37f71-cb02-4053-9e82-71a041cf8463","Type":"ContainerStarted","Data":"d27528a09e2f7e7a34ca84e63182ae0788361a2287cd0856c7a79968b1bd210e"} Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.296329 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.309744 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.327780 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.330657 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.332691 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.333081 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.343221 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.481638 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/84d2000a-d99f-4320-b178-874940ab7e9d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.481679 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqlxc\" (UniqueName: \"kubernetes.io/projected/84d2000a-d99f-4320-b178-874940ab7e9d-kube-api-access-zqlxc\") pod \"glance-default-internal-api-0\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.481708 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84d2000a-d99f-4320-b178-874940ab7e9d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.481771 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84d2000a-d99f-4320-b178-874940ab7e9d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.481886 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84d2000a-d99f-4320-b178-874940ab7e9d-logs\") pod \"glance-default-internal-api-0\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.482031 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84d2000a-d99f-4320-b178-874940ab7e9d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.482132 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/84d2000a-d99f-4320-b178-874940ab7e9d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.482214 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.583723 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84d2000a-d99f-4320-b178-874940ab7e9d-logs\") pod \"glance-default-internal-api-0\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.583782 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84d2000a-d99f-4320-b178-874940ab7e9d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.583806 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/84d2000a-d99f-4320-b178-874940ab7e9d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.583831 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.583906 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/84d2000a-d99f-4320-b178-874940ab7e9d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.583926 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqlxc\" (UniqueName: \"kubernetes.io/projected/84d2000a-d99f-4320-b178-874940ab7e9d-kube-api-access-zqlxc\") pod \"glance-default-internal-api-0\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.583946 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84d2000a-d99f-4320-b178-874940ab7e9d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.583988 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84d2000a-d99f-4320-b178-874940ab7e9d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.584223 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84d2000a-d99f-4320-b178-874940ab7e9d-logs\") pod \"glance-default-internal-api-0\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.584445 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/84d2000a-d99f-4320-b178-874940ab7e9d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.584711 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.595243 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/84d2000a-d99f-4320-b178-874940ab7e9d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.595308 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84d2000a-d99f-4320-b178-874940ab7e9d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.595994 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84d2000a-d99f-4320-b178-874940ab7e9d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.597135 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84d2000a-d99f-4320-b178-874940ab7e9d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.605700 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqlxc\" (UniqueName: \"kubernetes.io/projected/84d2000a-d99f-4320-b178-874940ab7e9d-kube-api-access-zqlxc\") pod \"glance-default-internal-api-0\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.626046 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " pod="openstack/glance-default-internal-api-0" Oct 02 10:08:17 crc kubenswrapper[4934]: I1002 10:08:17.654080 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 10:08:18 crc kubenswrapper[4934]: I1002 10:08:18.196923 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:08:18 crc kubenswrapper[4934]: I1002 10:08:18.270241 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"84d2000a-d99f-4320-b178-874940ab7e9d","Type":"ContainerStarted","Data":"5a03fdf3501c835a8877d0465aa7848d08e00f1009db66a43ba18f5aa66b3a00"} Oct 02 10:08:18 crc kubenswrapper[4934]: I1002 10:08:18.274904 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7cf37f71-cb02-4053-9e82-71a041cf8463","Type":"ContainerStarted","Data":"7736a402cca734bced39bc17ece6db88130e28e39cab46356279f01743005882"} Oct 02 10:08:18 crc kubenswrapper[4934]: I1002 10:08:18.930499 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e434dd5-17db-4f21-87d5-e41d09717e88" path="/var/lib/kubelet/pods/3e434dd5-17db-4f21-87d5-e41d09717e88/volumes" Oct 02 10:08:19 crc kubenswrapper[4934]: I1002 10:08:19.291457 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-57456ffd97-rqc8b" event={"ID":"6d79c694-2253-419e-b63b-4a38884ac57e","Type":"ContainerStarted","Data":"a68b3c26edd8aa7fbcd688303c8a83effaa456e62c9b48088363c448ed90aef6"} Oct 02 10:08:19 crc kubenswrapper[4934]: I1002 10:08:19.611722 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 02 10:08:19 crc kubenswrapper[4934]: I1002 10:08:19.951248 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:08:19 crc kubenswrapper[4934]: I1002 10:08:19.951848 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c3a1512e-c161-4264-ba36-148c10f52a36" containerName="ceilometer-central-agent" containerID="cri-o://41ac05160deb747befeacdc68c2f9f4a9b84267fe81e7a9f257a1cfbdd276ec2" gracePeriod=30 Oct 02 10:08:19 crc kubenswrapper[4934]: I1002 10:08:19.952434 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c3a1512e-c161-4264-ba36-148c10f52a36" containerName="proxy-httpd" containerID="cri-o://353a5f848560344ee82aa27e6d2845a5d78d06029b45a9ff5d536eda488efade" gracePeriod=30 Oct 02 10:08:19 crc kubenswrapper[4934]: I1002 10:08:19.952616 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c3a1512e-c161-4264-ba36-148c10f52a36" containerName="sg-core" containerID="cri-o://51def09dc4784bd05fa87ea6c929cfb1876b26f52ab5b324580b3d692ee0eb54" gracePeriod=30 Oct 02 10:08:19 crc kubenswrapper[4934]: I1002 10:08:19.952818 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c3a1512e-c161-4264-ba36-148c10f52a36" containerName="ceilometer-notification-agent" containerID="cri-o://01490ed90f471b4ad5147dc9eb7945067b7b6559aad2a7f47a8e593590da3985" gracePeriod=30 Oct 02 10:08:20 crc kubenswrapper[4934]: I1002 10:08:20.197718 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" Oct 02 10:08:20 crc kubenswrapper[4934]: I1002 10:08:20.329448 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79fb47bfff-98d57"] Oct 02 10:08:20 crc kubenswrapper[4934]: I1002 10:08:20.331214 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-79fb47bfff-98d57" podUID="f55848b6-dd77-46d8-8366-4715042cdd5d" containerName="dnsmasq-dns" containerID="cri-o://df78f82edde3a7888cee3d1396b9ac2314053be6d07b967766fd37331acc4de7" gracePeriod=10 Oct 02 10:08:20 crc kubenswrapper[4934]: I1002 10:08:20.367317 4934 generic.go:334] "Generic (PLEG): container finished" podID="c3a1512e-c161-4264-ba36-148c10f52a36" containerID="353a5f848560344ee82aa27e6d2845a5d78d06029b45a9ff5d536eda488efade" exitCode=0 Oct 02 10:08:20 crc kubenswrapper[4934]: I1002 10:08:20.367344 4934 generic.go:334] "Generic (PLEG): container finished" podID="c3a1512e-c161-4264-ba36-148c10f52a36" containerID="51def09dc4784bd05fa87ea6c929cfb1876b26f52ab5b324580b3d692ee0eb54" exitCode=2 Oct 02 10:08:20 crc kubenswrapper[4934]: I1002 10:08:20.367382 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3a1512e-c161-4264-ba36-148c10f52a36","Type":"ContainerDied","Data":"353a5f848560344ee82aa27e6d2845a5d78d06029b45a9ff5d536eda488efade"} Oct 02 10:08:20 crc kubenswrapper[4934]: I1002 10:08:20.367409 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3a1512e-c161-4264-ba36-148c10f52a36","Type":"ContainerDied","Data":"51def09dc4784bd05fa87ea6c929cfb1876b26f52ab5b324580b3d692ee0eb54"} Oct 02 10:08:20 crc kubenswrapper[4934]: I1002 10:08:20.403635 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-57456ffd97-rqc8b" event={"ID":"6d79c694-2253-419e-b63b-4a38884ac57e","Type":"ContainerStarted","Data":"959724f7a775ac28fa7ae269f3be79b301f1332c985d6f8a80a47def488f67f5"} Oct 02 10:08:20 crc kubenswrapper[4934]: I1002 10:08:20.404779 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-57456ffd97-rqc8b" Oct 02 10:08:20 crc kubenswrapper[4934]: I1002 10:08:20.412193 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"84d2000a-d99f-4320-b178-874940ab7e9d","Type":"ContainerStarted","Data":"9799cb2e3daec429b427c098af4fc69103ad295acfa71492709948a69984306d"} Oct 02 10:08:20 crc kubenswrapper[4934]: I1002 10:08:20.424899 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7cf37f71-cb02-4053-9e82-71a041cf8463","Type":"ContainerStarted","Data":"c22b0f0d6f101afda346dbc56f27da5e9351764d37ef615b51c1aba99b201822"} Oct 02 10:08:20 crc kubenswrapper[4934]: I1002 10:08:20.425050 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="7cf37f71-cb02-4053-9e82-71a041cf8463" containerName="glance-log" containerID="cri-o://7736a402cca734bced39bc17ece6db88130e28e39cab46356279f01743005882" gracePeriod=30 Oct 02 10:08:20 crc kubenswrapper[4934]: I1002 10:08:20.425458 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="7cf37f71-cb02-4053-9e82-71a041cf8463" containerName="glance-httpd" containerID="cri-o://c22b0f0d6f101afda346dbc56f27da5e9351764d37ef615b51c1aba99b201822" gracePeriod=30 Oct 02 10:08:20 crc kubenswrapper[4934]: I1002 10:08:20.429486 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-57456ffd97-rqc8b" podStartSLOduration=4.429461158 podStartE2EDuration="4.429461158s" podCreationTimestamp="2025-10-02 10:08:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:08:20.424611634 +0000 UTC m=+1172.177253156" watchObservedRunningTime="2025-10-02 10:08:20.429461158 +0000 UTC m=+1172.182102680" Oct 02 10:08:20 crc kubenswrapper[4934]: I1002 10:08:20.459360 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.459320052 podStartE2EDuration="5.459320052s" podCreationTimestamp="2025-10-02 10:08:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:08:20.447868396 +0000 UTC m=+1172.200509918" watchObservedRunningTime="2025-10-02 10:08:20.459320052 +0000 UTC m=+1172.211961574" Oct 02 10:08:20 crc kubenswrapper[4934]: I1002 10:08:20.831109 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79fb47bfff-98d57" Oct 02 10:08:20 crc kubenswrapper[4934]: I1002 10:08:20.962720 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-ovsdbserver-sb\") pod \"f55848b6-dd77-46d8-8366-4715042cdd5d\" (UID: \"f55848b6-dd77-46d8-8366-4715042cdd5d\") " Oct 02 10:08:20 crc kubenswrapper[4934]: I1002 10:08:20.962863 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-dns-swift-storage-0\") pod \"f55848b6-dd77-46d8-8366-4715042cdd5d\" (UID: \"f55848b6-dd77-46d8-8366-4715042cdd5d\") " Oct 02 10:08:20 crc kubenswrapper[4934]: I1002 10:08:20.963024 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-dns-svc\") pod \"f55848b6-dd77-46d8-8366-4715042cdd5d\" (UID: \"f55848b6-dd77-46d8-8366-4715042cdd5d\") " Oct 02 10:08:20 crc kubenswrapper[4934]: I1002 10:08:20.963084 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dc2vp\" (UniqueName: \"kubernetes.io/projected/f55848b6-dd77-46d8-8366-4715042cdd5d-kube-api-access-dc2vp\") pod \"f55848b6-dd77-46d8-8366-4715042cdd5d\" (UID: \"f55848b6-dd77-46d8-8366-4715042cdd5d\") " Oct 02 10:08:20 crc kubenswrapper[4934]: I1002 10:08:20.963122 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-config\") pod \"f55848b6-dd77-46d8-8366-4715042cdd5d\" (UID: \"f55848b6-dd77-46d8-8366-4715042cdd5d\") " Oct 02 10:08:20 crc kubenswrapper[4934]: I1002 10:08:20.963188 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-ovsdbserver-nb\") pod \"f55848b6-dd77-46d8-8366-4715042cdd5d\" (UID: \"f55848b6-dd77-46d8-8366-4715042cdd5d\") " Oct 02 10:08:20 crc kubenswrapper[4934]: I1002 10:08:20.976076 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f55848b6-dd77-46d8-8366-4715042cdd5d-kube-api-access-dc2vp" (OuterVolumeSpecName: "kube-api-access-dc2vp") pod "f55848b6-dd77-46d8-8366-4715042cdd5d" (UID: "f55848b6-dd77-46d8-8366-4715042cdd5d"). InnerVolumeSpecName "kube-api-access-dc2vp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.030127 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f55848b6-dd77-46d8-8366-4715042cdd5d" (UID: "f55848b6-dd77-46d8-8366-4715042cdd5d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.047025 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.048523 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f55848b6-dd77-46d8-8366-4715042cdd5d" (UID: "f55848b6-dd77-46d8-8366-4715042cdd5d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.049917 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f55848b6-dd77-46d8-8366-4715042cdd5d" (UID: "f55848b6-dd77-46d8-8366-4715042cdd5d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.056080 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f55848b6-dd77-46d8-8366-4715042cdd5d" (UID: "f55848b6-dd77-46d8-8366-4715042cdd5d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.059056 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-config" (OuterVolumeSpecName: "config") pod "f55848b6-dd77-46d8-8366-4715042cdd5d" (UID: "f55848b6-dd77-46d8-8366-4715042cdd5d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.067119 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.067152 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dc2vp\" (UniqueName: \"kubernetes.io/projected/f55848b6-dd77-46d8-8366-4715042cdd5d-kube-api-access-dc2vp\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.067161 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.067171 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.067181 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.067190 4934 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f55848b6-dd77-46d8-8366-4715042cdd5d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.168409 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cf37f71-cb02-4053-9e82-71a041cf8463-logs\") pod \"7cf37f71-cb02-4053-9e82-71a041cf8463\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") " Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.168585 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cf37f71-cb02-4053-9e82-71a041cf8463-combined-ca-bundle\") pod \"7cf37f71-cb02-4053-9e82-71a041cf8463\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") " Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.168645 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9sn4l\" (UniqueName: \"kubernetes.io/projected/7cf37f71-cb02-4053-9e82-71a041cf8463-kube-api-access-9sn4l\") pod \"7cf37f71-cb02-4053-9e82-71a041cf8463\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") " Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.168673 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cf37f71-cb02-4053-9e82-71a041cf8463-config-data\") pod \"7cf37f71-cb02-4053-9e82-71a041cf8463\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") " Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.168701 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7cf37f71-cb02-4053-9e82-71a041cf8463-httpd-run\") pod \"7cf37f71-cb02-4053-9e82-71a041cf8463\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") " Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.168761 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cf37f71-cb02-4053-9e82-71a041cf8463-scripts\") pod \"7cf37f71-cb02-4053-9e82-71a041cf8463\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") " Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.168786 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"7cf37f71-cb02-4053-9e82-71a041cf8463\" (UID: \"7cf37f71-cb02-4053-9e82-71a041cf8463\") " Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.169132 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cf37f71-cb02-4053-9e82-71a041cf8463-logs" (OuterVolumeSpecName: "logs") pod "7cf37f71-cb02-4053-9e82-71a041cf8463" (UID: "7cf37f71-cb02-4053-9e82-71a041cf8463"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.169335 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cf37f71-cb02-4053-9e82-71a041cf8463-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7cf37f71-cb02-4053-9e82-71a041cf8463" (UID: "7cf37f71-cb02-4053-9e82-71a041cf8463"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.173311 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cf37f71-cb02-4053-9e82-71a041cf8463-kube-api-access-9sn4l" (OuterVolumeSpecName: "kube-api-access-9sn4l") pod "7cf37f71-cb02-4053-9e82-71a041cf8463" (UID: "7cf37f71-cb02-4053-9e82-71a041cf8463"). InnerVolumeSpecName "kube-api-access-9sn4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.176451 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "7cf37f71-cb02-4053-9e82-71a041cf8463" (UID: "7cf37f71-cb02-4053-9e82-71a041cf8463"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.183319 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cf37f71-cb02-4053-9e82-71a041cf8463-scripts" (OuterVolumeSpecName: "scripts") pod "7cf37f71-cb02-4053-9e82-71a041cf8463" (UID: "7cf37f71-cb02-4053-9e82-71a041cf8463"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.199673 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cf37f71-cb02-4053-9e82-71a041cf8463-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7cf37f71-cb02-4053-9e82-71a041cf8463" (UID: "7cf37f71-cb02-4053-9e82-71a041cf8463"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.215778 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cf37f71-cb02-4053-9e82-71a041cf8463-config-data" (OuterVolumeSpecName: "config-data") pod "7cf37f71-cb02-4053-9e82-71a041cf8463" (UID: "7cf37f71-cb02-4053-9e82-71a041cf8463"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.270334 4934 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.270374 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cf37f71-cb02-4053-9e82-71a041cf8463-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.270385 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cf37f71-cb02-4053-9e82-71a041cf8463-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.270396 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9sn4l\" (UniqueName: \"kubernetes.io/projected/7cf37f71-cb02-4053-9e82-71a041cf8463-kube-api-access-9sn4l\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.270405 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cf37f71-cb02-4053-9e82-71a041cf8463-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.270412 4934 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7cf37f71-cb02-4053-9e82-71a041cf8463-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.270419 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cf37f71-cb02-4053-9e82-71a041cf8463-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.301991 4934 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.372264 4934 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.436184 4934 generic.go:334] "Generic (PLEG): container finished" podID="c3a1512e-c161-4264-ba36-148c10f52a36" containerID="01490ed90f471b4ad5147dc9eb7945067b7b6559aad2a7f47a8e593590da3985" exitCode=0 Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.436273 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3a1512e-c161-4264-ba36-148c10f52a36","Type":"ContainerDied","Data":"01490ed90f471b4ad5147dc9eb7945067b7b6559aad2a7f47a8e593590da3985"} Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.437998 4934 generic.go:334] "Generic (PLEG): container finished" podID="f55848b6-dd77-46d8-8366-4715042cdd5d" containerID="df78f82edde3a7888cee3d1396b9ac2314053be6d07b967766fd37331acc4de7" exitCode=0 Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.438080 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79fb47bfff-98d57" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.438084 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79fb47bfff-98d57" event={"ID":"f55848b6-dd77-46d8-8366-4715042cdd5d","Type":"ContainerDied","Data":"df78f82edde3a7888cee3d1396b9ac2314053be6d07b967766fd37331acc4de7"} Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.438127 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79fb47bfff-98d57" event={"ID":"f55848b6-dd77-46d8-8366-4715042cdd5d","Type":"ContainerDied","Data":"18645ce4ec2ad743abcf2b1feee55c89fe461a48841967d4b1116a5aefbce090"} Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.438150 4934 scope.go:117] "RemoveContainer" containerID="df78f82edde3a7888cee3d1396b9ac2314053be6d07b967766fd37331acc4de7" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.439836 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"84d2000a-d99f-4320-b178-874940ab7e9d","Type":"ContainerStarted","Data":"391ab06a0bf5924bac18ac518442b3a5539dd7de7e5001512639e7be11a54744"} Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.442793 4934 generic.go:334] "Generic (PLEG): container finished" podID="7cf37f71-cb02-4053-9e82-71a041cf8463" containerID="c22b0f0d6f101afda346dbc56f27da5e9351764d37ef615b51c1aba99b201822" exitCode=143 Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.442825 4934 generic.go:334] "Generic (PLEG): container finished" podID="7cf37f71-cb02-4053-9e82-71a041cf8463" containerID="7736a402cca734bced39bc17ece6db88130e28e39cab46356279f01743005882" exitCode=143 Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.443081 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7cf37f71-cb02-4053-9e82-71a041cf8463","Type":"ContainerDied","Data":"c22b0f0d6f101afda346dbc56f27da5e9351764d37ef615b51c1aba99b201822"} Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.443094 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.443120 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7cf37f71-cb02-4053-9e82-71a041cf8463","Type":"ContainerDied","Data":"7736a402cca734bced39bc17ece6db88130e28e39cab46356279f01743005882"} Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.443131 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7cf37f71-cb02-4053-9e82-71a041cf8463","Type":"ContainerDied","Data":"d27528a09e2f7e7a34ca84e63182ae0788361a2287cd0856c7a79968b1bd210e"} Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.470860 4934 scope.go:117] "RemoveContainer" containerID="79c57afcd39582909d2f4e613b9d5c29d54192176c3f8f3aaa960f62fd72a9ee" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.475472 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.475450094 podStartE2EDuration="4.475450094s" podCreationTimestamp="2025-10-02 10:08:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:08:21.471178036 +0000 UTC m=+1173.223819558" watchObservedRunningTime="2025-10-02 10:08:21.475450094 +0000 UTC m=+1173.228091616" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.495747 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79fb47bfff-98d57"] Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.503073 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79fb47bfff-98d57"] Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.521610 4934 scope.go:117] "RemoveContainer" containerID="df78f82edde3a7888cee3d1396b9ac2314053be6d07b967766fd37331acc4de7" Oct 02 10:08:21 crc kubenswrapper[4934]: E1002 10:08:21.525717 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df78f82edde3a7888cee3d1396b9ac2314053be6d07b967766fd37331acc4de7\": container with ID starting with df78f82edde3a7888cee3d1396b9ac2314053be6d07b967766fd37331acc4de7 not found: ID does not exist" containerID="df78f82edde3a7888cee3d1396b9ac2314053be6d07b967766fd37331acc4de7" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.525769 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df78f82edde3a7888cee3d1396b9ac2314053be6d07b967766fd37331acc4de7"} err="failed to get container status \"df78f82edde3a7888cee3d1396b9ac2314053be6d07b967766fd37331acc4de7\": rpc error: code = NotFound desc = could not find container \"df78f82edde3a7888cee3d1396b9ac2314053be6d07b967766fd37331acc4de7\": container with ID starting with df78f82edde3a7888cee3d1396b9ac2314053be6d07b967766fd37331acc4de7 not found: ID does not exist" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.525806 4934 scope.go:117] "RemoveContainer" containerID="79c57afcd39582909d2f4e613b9d5c29d54192176c3f8f3aaa960f62fd72a9ee" Oct 02 10:08:21 crc kubenswrapper[4934]: E1002 10:08:21.526560 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79c57afcd39582909d2f4e613b9d5c29d54192176c3f8f3aaa960f62fd72a9ee\": container with ID starting with 79c57afcd39582909d2f4e613b9d5c29d54192176c3f8f3aaa960f62fd72a9ee not found: ID does not exist" containerID="79c57afcd39582909d2f4e613b9d5c29d54192176c3f8f3aaa960f62fd72a9ee" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.526642 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79c57afcd39582909d2f4e613b9d5c29d54192176c3f8f3aaa960f62fd72a9ee"} err="failed to get container status \"79c57afcd39582909d2f4e613b9d5c29d54192176c3f8f3aaa960f62fd72a9ee\": rpc error: code = NotFound desc = could not find container \"79c57afcd39582909d2f4e613b9d5c29d54192176c3f8f3aaa960f62fd72a9ee\": container with ID starting with 79c57afcd39582909d2f4e613b9d5c29d54192176c3f8f3aaa960f62fd72a9ee not found: ID does not exist" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.526662 4934 scope.go:117] "RemoveContainer" containerID="c22b0f0d6f101afda346dbc56f27da5e9351764d37ef615b51c1aba99b201822" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.526817 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.551414 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.568402 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:08:21 crc kubenswrapper[4934]: E1002 10:08:21.568948 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f55848b6-dd77-46d8-8366-4715042cdd5d" containerName="init" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.568972 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f55848b6-dd77-46d8-8366-4715042cdd5d" containerName="init" Oct 02 10:08:21 crc kubenswrapper[4934]: E1002 10:08:21.569006 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cf37f71-cb02-4053-9e82-71a041cf8463" containerName="glance-httpd" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.569014 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cf37f71-cb02-4053-9e82-71a041cf8463" containerName="glance-httpd" Oct 02 10:08:21 crc kubenswrapper[4934]: E1002 10:08:21.569056 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cf37f71-cb02-4053-9e82-71a041cf8463" containerName="glance-log" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.569064 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cf37f71-cb02-4053-9e82-71a041cf8463" containerName="glance-log" Oct 02 10:08:21 crc kubenswrapper[4934]: E1002 10:08:21.569079 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f55848b6-dd77-46d8-8366-4715042cdd5d" containerName="dnsmasq-dns" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.569091 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f55848b6-dd77-46d8-8366-4715042cdd5d" containerName="dnsmasq-dns" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.569298 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cf37f71-cb02-4053-9e82-71a041cf8463" containerName="glance-httpd" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.569319 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cf37f71-cb02-4053-9e82-71a041cf8463" containerName="glance-log" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.569333 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f55848b6-dd77-46d8-8366-4715042cdd5d" containerName="dnsmasq-dns" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.570481 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.579079 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.579083 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.586978 4934 scope.go:117] "RemoveContainer" containerID="7736a402cca734bced39bc17ece6db88130e28e39cab46356279f01743005882" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.634696 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.652185 4934 scope.go:117] "RemoveContainer" containerID="c22b0f0d6f101afda346dbc56f27da5e9351764d37ef615b51c1aba99b201822" Oct 02 10:08:21 crc kubenswrapper[4934]: E1002 10:08:21.652662 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c22b0f0d6f101afda346dbc56f27da5e9351764d37ef615b51c1aba99b201822\": container with ID starting with c22b0f0d6f101afda346dbc56f27da5e9351764d37ef615b51c1aba99b201822 not found: ID does not exist" containerID="c22b0f0d6f101afda346dbc56f27da5e9351764d37ef615b51c1aba99b201822" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.652723 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c22b0f0d6f101afda346dbc56f27da5e9351764d37ef615b51c1aba99b201822"} err="failed to get container status \"c22b0f0d6f101afda346dbc56f27da5e9351764d37ef615b51c1aba99b201822\": rpc error: code = NotFound desc = could not find container \"c22b0f0d6f101afda346dbc56f27da5e9351764d37ef615b51c1aba99b201822\": container with ID starting with c22b0f0d6f101afda346dbc56f27da5e9351764d37ef615b51c1aba99b201822 not found: ID does not exist" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.652770 4934 scope.go:117] "RemoveContainer" containerID="7736a402cca734bced39bc17ece6db88130e28e39cab46356279f01743005882" Oct 02 10:08:21 crc kubenswrapper[4934]: E1002 10:08:21.656849 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7736a402cca734bced39bc17ece6db88130e28e39cab46356279f01743005882\": container with ID starting with 7736a402cca734bced39bc17ece6db88130e28e39cab46356279f01743005882 not found: ID does not exist" containerID="7736a402cca734bced39bc17ece6db88130e28e39cab46356279f01743005882" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.656882 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7736a402cca734bced39bc17ece6db88130e28e39cab46356279f01743005882"} err="failed to get container status \"7736a402cca734bced39bc17ece6db88130e28e39cab46356279f01743005882\": rpc error: code = NotFound desc = could not find container \"7736a402cca734bced39bc17ece6db88130e28e39cab46356279f01743005882\": container with ID starting with 7736a402cca734bced39bc17ece6db88130e28e39cab46356279f01743005882 not found: ID does not exist" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.656902 4934 scope.go:117] "RemoveContainer" containerID="c22b0f0d6f101afda346dbc56f27da5e9351764d37ef615b51c1aba99b201822" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.659864 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c22b0f0d6f101afda346dbc56f27da5e9351764d37ef615b51c1aba99b201822"} err="failed to get container status \"c22b0f0d6f101afda346dbc56f27da5e9351764d37ef615b51c1aba99b201822\": rpc error: code = NotFound desc = could not find container \"c22b0f0d6f101afda346dbc56f27da5e9351764d37ef615b51c1aba99b201822\": container with ID starting with c22b0f0d6f101afda346dbc56f27da5e9351764d37ef615b51c1aba99b201822 not found: ID does not exist" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.659909 4934 scope.go:117] "RemoveContainer" containerID="7736a402cca734bced39bc17ece6db88130e28e39cab46356279f01743005882" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.660271 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7736a402cca734bced39bc17ece6db88130e28e39cab46356279f01743005882"} err="failed to get container status \"7736a402cca734bced39bc17ece6db88130e28e39cab46356279f01743005882\": rpc error: code = NotFound desc = could not find container \"7736a402cca734bced39bc17ece6db88130e28e39cab46356279f01743005882\": container with ID starting with 7736a402cca734bced39bc17ece6db88130e28e39cab46356279f01743005882 not found: ID does not exist" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.683923 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9165cca6-da95-45ae-87c3-b2829756db3b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.684489 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9165cca6-da95-45ae-87c3-b2829756db3b-config-data\") pod \"glance-default-external-api-0\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.684614 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9165cca6-da95-45ae-87c3-b2829756db3b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.684652 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9165cca6-da95-45ae-87c3-b2829756db3b-logs\") pod \"glance-default-external-api-0\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.684682 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.684753 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qz6fh\" (UniqueName: \"kubernetes.io/projected/9165cca6-da95-45ae-87c3-b2829756db3b-kube-api-access-qz6fh\") pod \"glance-default-external-api-0\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.684816 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9165cca6-da95-45ae-87c3-b2829756db3b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.684854 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9165cca6-da95-45ae-87c3-b2829756db3b-scripts\") pod \"glance-default-external-api-0\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.786286 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9165cca6-da95-45ae-87c3-b2829756db3b-config-data\") pod \"glance-default-external-api-0\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.786354 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9165cca6-da95-45ae-87c3-b2829756db3b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.786378 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9165cca6-da95-45ae-87c3-b2829756db3b-logs\") pod \"glance-default-external-api-0\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.786397 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.786785 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.787057 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9165cca6-da95-45ae-87c3-b2829756db3b-logs\") pod \"glance-default-external-api-0\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.790303 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qz6fh\" (UniqueName: \"kubernetes.io/projected/9165cca6-da95-45ae-87c3-b2829756db3b-kube-api-access-qz6fh\") pod \"glance-default-external-api-0\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.790419 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9165cca6-da95-45ae-87c3-b2829756db3b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.790443 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9165cca6-da95-45ae-87c3-b2829756db3b-scripts\") pod \"glance-default-external-api-0\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.791608 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9165cca6-da95-45ae-87c3-b2829756db3b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.793120 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9165cca6-da95-45ae-87c3-b2829756db3b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.793872 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9165cca6-da95-45ae-87c3-b2829756db3b-config-data\") pod \"glance-default-external-api-0\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.794364 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9165cca6-da95-45ae-87c3-b2829756db3b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.795663 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9165cca6-da95-45ae-87c3-b2829756db3b-scripts\") pod \"glance-default-external-api-0\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.801920 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9165cca6-da95-45ae-87c3-b2829756db3b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.807237 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qz6fh\" (UniqueName: \"kubernetes.io/projected/9165cca6-da95-45ae-87c3-b2829756db3b-kube-api-access-qz6fh\") pod \"glance-default-external-api-0\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.825682 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " pod="openstack/glance-default-external-api-0" Oct 02 10:08:21 crc kubenswrapper[4934]: I1002 10:08:21.929025 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 10:08:22 crc kubenswrapper[4934]: I1002 10:08:22.559273 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:08:22 crc kubenswrapper[4934]: I1002 10:08:22.934201 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cf37f71-cb02-4053-9e82-71a041cf8463" path="/var/lib/kubelet/pods/7cf37f71-cb02-4053-9e82-71a041cf8463/volumes" Oct 02 10:08:22 crc kubenswrapper[4934]: I1002 10:08:22.935281 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f55848b6-dd77-46d8-8366-4715042cdd5d" path="/var/lib/kubelet/pods/f55848b6-dd77-46d8-8366-4715042cdd5d/volumes" Oct 02 10:08:23 crc kubenswrapper[4934]: I1002 10:08:23.476998 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9165cca6-da95-45ae-87c3-b2829756db3b","Type":"ContainerStarted","Data":"9dde5aaa3e65635d53137792cb4e654aa61c584c4bc197377b96acd4ff67091e"} Oct 02 10:08:23 crc kubenswrapper[4934]: I1002 10:08:23.477289 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9165cca6-da95-45ae-87c3-b2829756db3b","Type":"ContainerStarted","Data":"76d825cb9e54693e07042d9c95aebb56e6bc394a03bc211dedb6af4cb0844a87"} Oct 02 10:08:24 crc kubenswrapper[4934]: I1002 10:08:24.488963 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9165cca6-da95-45ae-87c3-b2829756db3b","Type":"ContainerStarted","Data":"e38759eb2a2651b29d3a6703a0da0365f7f7843aad2b4f894f68ed24320e6d21"} Oct 02 10:08:24 crc kubenswrapper[4934]: I1002 10:08:24.592685 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.592669068 podStartE2EDuration="3.592669068s" podCreationTimestamp="2025-10-02 10:08:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:08:24.523265782 +0000 UTC m=+1176.275907294" watchObservedRunningTime="2025-10-02 10:08:24.592669068 +0000 UTC m=+1176.345310600" Oct 02 10:08:24 crc kubenswrapper[4934]: I1002 10:08:24.597018 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-4z9rv"] Oct 02 10:08:24 crc kubenswrapper[4934]: I1002 10:08:24.598162 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4z9rv" Oct 02 10:08:24 crc kubenswrapper[4934]: I1002 10:08:24.608624 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-4z9rv"] Oct 02 10:08:24 crc kubenswrapper[4934]: I1002 10:08:24.658402 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc7px\" (UniqueName: \"kubernetes.io/projected/f57c470b-0a2a-4341-ae16-6d5ac848ade0-kube-api-access-dc7px\") pod \"nova-api-db-create-4z9rv\" (UID: \"f57c470b-0a2a-4341-ae16-6d5ac848ade0\") " pod="openstack/nova-api-db-create-4z9rv" Oct 02 10:08:24 crc kubenswrapper[4934]: I1002 10:08:24.696520 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-h84gj"] Oct 02 10:08:24 crc kubenswrapper[4934]: I1002 10:08:24.698054 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-h84gj" Oct 02 10:08:24 crc kubenswrapper[4934]: I1002 10:08:24.712914 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-h84gj"] Oct 02 10:08:24 crc kubenswrapper[4934]: I1002 10:08:24.759678 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qkxx\" (UniqueName: \"kubernetes.io/projected/26f3aea9-8e30-4ec3-b440-c00fe8f624da-kube-api-access-7qkxx\") pod \"nova-cell0-db-create-h84gj\" (UID: \"26f3aea9-8e30-4ec3-b440-c00fe8f624da\") " pod="openstack/nova-cell0-db-create-h84gj" Oct 02 10:08:24 crc kubenswrapper[4934]: I1002 10:08:24.760028 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc7px\" (UniqueName: \"kubernetes.io/projected/f57c470b-0a2a-4341-ae16-6d5ac848ade0-kube-api-access-dc7px\") pod \"nova-api-db-create-4z9rv\" (UID: \"f57c470b-0a2a-4341-ae16-6d5ac848ade0\") " pod="openstack/nova-api-db-create-4z9rv" Oct 02 10:08:24 crc kubenswrapper[4934]: I1002 10:08:24.784603 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc7px\" (UniqueName: \"kubernetes.io/projected/f57c470b-0a2a-4341-ae16-6d5ac848ade0-kube-api-access-dc7px\") pod \"nova-api-db-create-4z9rv\" (UID: \"f57c470b-0a2a-4341-ae16-6d5ac848ade0\") " pod="openstack/nova-api-db-create-4z9rv" Oct 02 10:08:24 crc kubenswrapper[4934]: I1002 10:08:24.861951 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qkxx\" (UniqueName: \"kubernetes.io/projected/26f3aea9-8e30-4ec3-b440-c00fe8f624da-kube-api-access-7qkxx\") pod \"nova-cell0-db-create-h84gj\" (UID: \"26f3aea9-8e30-4ec3-b440-c00fe8f624da\") " pod="openstack/nova-cell0-db-create-h84gj" Oct 02 10:08:24 crc kubenswrapper[4934]: I1002 10:08:24.891778 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qkxx\" (UniqueName: \"kubernetes.io/projected/26f3aea9-8e30-4ec3-b440-c00fe8f624da-kube-api-access-7qkxx\") pod \"nova-cell0-db-create-h84gj\" (UID: \"26f3aea9-8e30-4ec3-b440-c00fe8f624da\") " pod="openstack/nova-cell0-db-create-h84gj" Oct 02 10:08:24 crc kubenswrapper[4934]: I1002 10:08:24.917978 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4z9rv" Oct 02 10:08:24 crc kubenswrapper[4934]: I1002 10:08:24.933118 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-kh9px"] Oct 02 10:08:24 crc kubenswrapper[4934]: I1002 10:08:24.936074 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kh9px" Oct 02 10:08:24 crc kubenswrapper[4934]: I1002 10:08:24.965840 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-kh9px"] Oct 02 10:08:25 crc kubenswrapper[4934]: I1002 10:08:25.008941 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvn4d\" (UniqueName: \"kubernetes.io/projected/09b1ea6c-4d4a-427d-a66e-e3db8b198316-kube-api-access-zvn4d\") pod \"nova-cell1-db-create-kh9px\" (UID: \"09b1ea6c-4d4a-427d-a66e-e3db8b198316\") " pod="openstack/nova-cell1-db-create-kh9px" Oct 02 10:08:25 crc kubenswrapper[4934]: I1002 10:08:25.028980 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-h84gj" Oct 02 10:08:25 crc kubenswrapper[4934]: I1002 10:08:25.132735 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvn4d\" (UniqueName: \"kubernetes.io/projected/09b1ea6c-4d4a-427d-a66e-e3db8b198316-kube-api-access-zvn4d\") pod \"nova-cell1-db-create-kh9px\" (UID: \"09b1ea6c-4d4a-427d-a66e-e3db8b198316\") " pod="openstack/nova-cell1-db-create-kh9px" Oct 02 10:08:25 crc kubenswrapper[4934]: I1002 10:08:25.150341 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvn4d\" (UniqueName: \"kubernetes.io/projected/09b1ea6c-4d4a-427d-a66e-e3db8b198316-kube-api-access-zvn4d\") pod \"nova-cell1-db-create-kh9px\" (UID: \"09b1ea6c-4d4a-427d-a66e-e3db8b198316\") " pod="openstack/nova-cell1-db-create-kh9px" Oct 02 10:08:25 crc kubenswrapper[4934]: I1002 10:08:25.392258 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kh9px" Oct 02 10:08:25 crc kubenswrapper[4934]: I1002 10:08:25.440898 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-4z9rv"] Oct 02 10:08:25 crc kubenswrapper[4934]: I1002 10:08:25.541865 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4z9rv" event={"ID":"f57c470b-0a2a-4341-ae16-6d5ac848ade0","Type":"ContainerStarted","Data":"98c96cadec8fc9f74eb7f134ce267562f51435dd24afc8591c36ed501d992fa8"} Oct 02 10:08:25 crc kubenswrapper[4934]: I1002 10:08:25.566291 4934 generic.go:334] "Generic (PLEG): container finished" podID="c3a1512e-c161-4264-ba36-148c10f52a36" containerID="41ac05160deb747befeacdc68c2f9f4a9b84267fe81e7a9f257a1cfbdd276ec2" exitCode=0 Oct 02 10:08:25 crc kubenswrapper[4934]: I1002 10:08:25.566364 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3a1512e-c161-4264-ba36-148c10f52a36","Type":"ContainerDied","Data":"41ac05160deb747befeacdc68c2f9f4a9b84267fe81e7a9f257a1cfbdd276ec2"} Oct 02 10:08:25 crc kubenswrapper[4934]: I1002 10:08:25.567245 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-h84gj"] Oct 02 10:08:25 crc kubenswrapper[4934]: I1002 10:08:25.780888 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:08:25 crc kubenswrapper[4934]: I1002 10:08:25.941499 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-kh9px"] Oct 02 10:08:25 crc kubenswrapper[4934]: I1002 10:08:25.949435 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3a1512e-c161-4264-ba36-148c10f52a36-scripts\") pod \"c3a1512e-c161-4264-ba36-148c10f52a36\" (UID: \"c3a1512e-c161-4264-ba36-148c10f52a36\") " Oct 02 10:08:25 crc kubenswrapper[4934]: I1002 10:08:25.949549 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3a1512e-c161-4264-ba36-148c10f52a36-run-httpd\") pod \"c3a1512e-c161-4264-ba36-148c10f52a36\" (UID: \"c3a1512e-c161-4264-ba36-148c10f52a36\") " Oct 02 10:08:25 crc kubenswrapper[4934]: I1002 10:08:25.949586 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3a1512e-c161-4264-ba36-148c10f52a36-log-httpd\") pod \"c3a1512e-c161-4264-ba36-148c10f52a36\" (UID: \"c3a1512e-c161-4264-ba36-148c10f52a36\") " Oct 02 10:08:25 crc kubenswrapper[4934]: I1002 10:08:25.949620 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2zxvn\" (UniqueName: \"kubernetes.io/projected/c3a1512e-c161-4264-ba36-148c10f52a36-kube-api-access-2zxvn\") pod \"c3a1512e-c161-4264-ba36-148c10f52a36\" (UID: \"c3a1512e-c161-4264-ba36-148c10f52a36\") " Oct 02 10:08:25 crc kubenswrapper[4934]: I1002 10:08:25.949652 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3a1512e-c161-4264-ba36-148c10f52a36-config-data\") pod \"c3a1512e-c161-4264-ba36-148c10f52a36\" (UID: \"c3a1512e-c161-4264-ba36-148c10f52a36\") " Oct 02 10:08:25 crc kubenswrapper[4934]: I1002 10:08:25.949704 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c3a1512e-c161-4264-ba36-148c10f52a36-sg-core-conf-yaml\") pod \"c3a1512e-c161-4264-ba36-148c10f52a36\" (UID: \"c3a1512e-c161-4264-ba36-148c10f52a36\") " Oct 02 10:08:25 crc kubenswrapper[4934]: I1002 10:08:25.949836 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3a1512e-c161-4264-ba36-148c10f52a36-combined-ca-bundle\") pod \"c3a1512e-c161-4264-ba36-148c10f52a36\" (UID: \"c3a1512e-c161-4264-ba36-148c10f52a36\") " Oct 02 10:08:25 crc kubenswrapper[4934]: I1002 10:08:25.950115 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3a1512e-c161-4264-ba36-148c10f52a36-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c3a1512e-c161-4264-ba36-148c10f52a36" (UID: "c3a1512e-c161-4264-ba36-148c10f52a36"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:08:25 crc kubenswrapper[4934]: I1002 10:08:25.950414 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3a1512e-c161-4264-ba36-148c10f52a36-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c3a1512e-c161-4264-ba36-148c10f52a36" (UID: "c3a1512e-c161-4264-ba36-148c10f52a36"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:08:25 crc kubenswrapper[4934]: I1002 10:08:25.955780 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3a1512e-c161-4264-ba36-148c10f52a36-kube-api-access-2zxvn" (OuterVolumeSpecName: "kube-api-access-2zxvn") pod "c3a1512e-c161-4264-ba36-148c10f52a36" (UID: "c3a1512e-c161-4264-ba36-148c10f52a36"). InnerVolumeSpecName "kube-api-access-2zxvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:08:25 crc kubenswrapper[4934]: I1002 10:08:25.960545 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3a1512e-c161-4264-ba36-148c10f52a36-scripts" (OuterVolumeSpecName: "scripts") pod "c3a1512e-c161-4264-ba36-148c10f52a36" (UID: "c3a1512e-c161-4264-ba36-148c10f52a36"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:25 crc kubenswrapper[4934]: I1002 10:08:25.995466 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3a1512e-c161-4264-ba36-148c10f52a36-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c3a1512e-c161-4264-ba36-148c10f52a36" (UID: "c3a1512e-c161-4264-ba36-148c10f52a36"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.052489 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3a1512e-c161-4264-ba36-148c10f52a36-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.052739 4934 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3a1512e-c161-4264-ba36-148c10f52a36-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.052853 4934 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3a1512e-c161-4264-ba36-148c10f52a36-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.052967 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2zxvn\" (UniqueName: \"kubernetes.io/projected/c3a1512e-c161-4264-ba36-148c10f52a36-kube-api-access-2zxvn\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.053059 4934 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c3a1512e-c161-4264-ba36-148c10f52a36-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.061789 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3a1512e-c161-4264-ba36-148c10f52a36-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c3a1512e-c161-4264-ba36-148c10f52a36" (UID: "c3a1512e-c161-4264-ba36-148c10f52a36"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.099710 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3a1512e-c161-4264-ba36-148c10f52a36-config-data" (OuterVolumeSpecName: "config-data") pod "c3a1512e-c161-4264-ba36-148c10f52a36" (UID: "c3a1512e-c161-4264-ba36-148c10f52a36"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.154784 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3a1512e-c161-4264-ba36-148c10f52a36-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.154825 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3a1512e-c161-4264-ba36-148c10f52a36-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.578328 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3a1512e-c161-4264-ba36-148c10f52a36","Type":"ContainerDied","Data":"553f83e6f2e0bb757bb167a63920f7ca1f53ee05888db0dd64a03851c737b6ed"} Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.578710 4934 scope.go:117] "RemoveContainer" containerID="353a5f848560344ee82aa27e6d2845a5d78d06029b45a9ff5d536eda488efade" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.578605 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.582458 4934 generic.go:334] "Generic (PLEG): container finished" podID="09b1ea6c-4d4a-427d-a66e-e3db8b198316" containerID="e026663f32d10778f738428a7a66f6498bce90ad6030d9b02bb84a6b4e5df0f0" exitCode=0 Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.582517 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-kh9px" event={"ID":"09b1ea6c-4d4a-427d-a66e-e3db8b198316","Type":"ContainerDied","Data":"e026663f32d10778f738428a7a66f6498bce90ad6030d9b02bb84a6b4e5df0f0"} Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.582542 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-kh9px" event={"ID":"09b1ea6c-4d4a-427d-a66e-e3db8b198316","Type":"ContainerStarted","Data":"1ae6fabae2fa27df456226420f3311d2a3908de63e7a14c5ee520811a70e98c7"} Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.584542 4934 generic.go:334] "Generic (PLEG): container finished" podID="26f3aea9-8e30-4ec3-b440-c00fe8f624da" containerID="4b29478d61836822ab583964f5cda288f9e4166121ace0a26a374d08d5e1860a" exitCode=0 Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.584633 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-h84gj" event={"ID":"26f3aea9-8e30-4ec3-b440-c00fe8f624da","Type":"ContainerDied","Data":"4b29478d61836822ab583964f5cda288f9e4166121ace0a26a374d08d5e1860a"} Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.584651 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-h84gj" event={"ID":"26f3aea9-8e30-4ec3-b440-c00fe8f624da","Type":"ContainerStarted","Data":"7d4c5613a1af27fbd439a142db3cfcb83fb8da809b02d31244b45b9418034766"} Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.589040 4934 generic.go:334] "Generic (PLEG): container finished" podID="f57c470b-0a2a-4341-ae16-6d5ac848ade0" containerID="2f94b6c5f263f4184b746c3ade707cb59c40c51fe029d679e0ac873fb6c628e2" exitCode=0 Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.589074 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4z9rv" event={"ID":"f57c470b-0a2a-4341-ae16-6d5ac848ade0","Type":"ContainerDied","Data":"2f94b6c5f263f4184b746c3ade707cb59c40c51fe029d679e0ac873fb6c628e2"} Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.607577 4934 scope.go:117] "RemoveContainer" containerID="51def09dc4784bd05fa87ea6c929cfb1876b26f52ab5b324580b3d692ee0eb54" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.645775 4934 scope.go:117] "RemoveContainer" containerID="01490ed90f471b4ad5147dc9eb7945067b7b6559aad2a7f47a8e593590da3985" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.651255 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.663413 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.686250 4934 scope.go:117] "RemoveContainer" containerID="41ac05160deb747befeacdc68c2f9f4a9b84267fe81e7a9f257a1cfbdd276ec2" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.695171 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:08:26 crc kubenswrapper[4934]: E1002 10:08:26.695612 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3a1512e-c161-4264-ba36-148c10f52a36" containerName="proxy-httpd" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.695625 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3a1512e-c161-4264-ba36-148c10f52a36" containerName="proxy-httpd" Oct 02 10:08:26 crc kubenswrapper[4934]: E1002 10:08:26.695645 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3a1512e-c161-4264-ba36-148c10f52a36" containerName="sg-core" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.695651 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3a1512e-c161-4264-ba36-148c10f52a36" containerName="sg-core" Oct 02 10:08:26 crc kubenswrapper[4934]: E1002 10:08:26.695666 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3a1512e-c161-4264-ba36-148c10f52a36" containerName="ceilometer-notification-agent" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.695672 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3a1512e-c161-4264-ba36-148c10f52a36" containerName="ceilometer-notification-agent" Oct 02 10:08:26 crc kubenswrapper[4934]: E1002 10:08:26.695688 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3a1512e-c161-4264-ba36-148c10f52a36" containerName="ceilometer-central-agent" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.695697 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3a1512e-c161-4264-ba36-148c10f52a36" containerName="ceilometer-central-agent" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.695869 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3a1512e-c161-4264-ba36-148c10f52a36" containerName="ceilometer-notification-agent" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.695893 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3a1512e-c161-4264-ba36-148c10f52a36" containerName="proxy-httpd" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.695913 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3a1512e-c161-4264-ba36-148c10f52a36" containerName="ceilometer-central-agent" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.695926 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3a1512e-c161-4264-ba36-148c10f52a36" containerName="sg-core" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.697592 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.700079 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.700229 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.715802 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.869432 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7efebc3a-dab4-4039-8fe1-e2e361666b97-scripts\") pod \"ceilometer-0\" (UID: \"7efebc3a-dab4-4039-8fe1-e2e361666b97\") " pod="openstack/ceilometer-0" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.869504 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7efebc3a-dab4-4039-8fe1-e2e361666b97-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7efebc3a-dab4-4039-8fe1-e2e361666b97\") " pod="openstack/ceilometer-0" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.869694 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4w8l\" (UniqueName: \"kubernetes.io/projected/7efebc3a-dab4-4039-8fe1-e2e361666b97-kube-api-access-p4w8l\") pod \"ceilometer-0\" (UID: \"7efebc3a-dab4-4039-8fe1-e2e361666b97\") " pod="openstack/ceilometer-0" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.870049 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7efebc3a-dab4-4039-8fe1-e2e361666b97-config-data\") pod \"ceilometer-0\" (UID: \"7efebc3a-dab4-4039-8fe1-e2e361666b97\") " pod="openstack/ceilometer-0" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.870089 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7efebc3a-dab4-4039-8fe1-e2e361666b97-log-httpd\") pod \"ceilometer-0\" (UID: \"7efebc3a-dab4-4039-8fe1-e2e361666b97\") " pod="openstack/ceilometer-0" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.870117 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7efebc3a-dab4-4039-8fe1-e2e361666b97-run-httpd\") pod \"ceilometer-0\" (UID: \"7efebc3a-dab4-4039-8fe1-e2e361666b97\") " pod="openstack/ceilometer-0" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.870232 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7efebc3a-dab4-4039-8fe1-e2e361666b97-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7efebc3a-dab4-4039-8fe1-e2e361666b97\") " pod="openstack/ceilometer-0" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.922982 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3a1512e-c161-4264-ba36-148c10f52a36" path="/var/lib/kubelet/pods/c3a1512e-c161-4264-ba36-148c10f52a36/volumes" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.971261 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7efebc3a-dab4-4039-8fe1-e2e361666b97-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7efebc3a-dab4-4039-8fe1-e2e361666b97\") " pod="openstack/ceilometer-0" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.971315 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4w8l\" (UniqueName: \"kubernetes.io/projected/7efebc3a-dab4-4039-8fe1-e2e361666b97-kube-api-access-p4w8l\") pod \"ceilometer-0\" (UID: \"7efebc3a-dab4-4039-8fe1-e2e361666b97\") " pod="openstack/ceilometer-0" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.971393 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7efebc3a-dab4-4039-8fe1-e2e361666b97-config-data\") pod \"ceilometer-0\" (UID: \"7efebc3a-dab4-4039-8fe1-e2e361666b97\") " pod="openstack/ceilometer-0" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.971412 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7efebc3a-dab4-4039-8fe1-e2e361666b97-log-httpd\") pod \"ceilometer-0\" (UID: \"7efebc3a-dab4-4039-8fe1-e2e361666b97\") " pod="openstack/ceilometer-0" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.971431 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7efebc3a-dab4-4039-8fe1-e2e361666b97-run-httpd\") pod \"ceilometer-0\" (UID: \"7efebc3a-dab4-4039-8fe1-e2e361666b97\") " pod="openstack/ceilometer-0" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.971462 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7efebc3a-dab4-4039-8fe1-e2e361666b97-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7efebc3a-dab4-4039-8fe1-e2e361666b97\") " pod="openstack/ceilometer-0" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.971512 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7efebc3a-dab4-4039-8fe1-e2e361666b97-scripts\") pod \"ceilometer-0\" (UID: \"7efebc3a-dab4-4039-8fe1-e2e361666b97\") " pod="openstack/ceilometer-0" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.976763 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7efebc3a-dab4-4039-8fe1-e2e361666b97-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7efebc3a-dab4-4039-8fe1-e2e361666b97\") " pod="openstack/ceilometer-0" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.976826 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7efebc3a-dab4-4039-8fe1-e2e361666b97-run-httpd\") pod \"ceilometer-0\" (UID: \"7efebc3a-dab4-4039-8fe1-e2e361666b97\") " pod="openstack/ceilometer-0" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.977036 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7efebc3a-dab4-4039-8fe1-e2e361666b97-log-httpd\") pod \"ceilometer-0\" (UID: \"7efebc3a-dab4-4039-8fe1-e2e361666b97\") " pod="openstack/ceilometer-0" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.980691 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7efebc3a-dab4-4039-8fe1-e2e361666b97-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7efebc3a-dab4-4039-8fe1-e2e361666b97\") " pod="openstack/ceilometer-0" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.985849 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7efebc3a-dab4-4039-8fe1-e2e361666b97-config-data\") pod \"ceilometer-0\" (UID: \"7efebc3a-dab4-4039-8fe1-e2e361666b97\") " pod="openstack/ceilometer-0" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.988786 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7efebc3a-dab4-4039-8fe1-e2e361666b97-scripts\") pod \"ceilometer-0\" (UID: \"7efebc3a-dab4-4039-8fe1-e2e361666b97\") " pod="openstack/ceilometer-0" Oct 02 10:08:26 crc kubenswrapper[4934]: I1002 10:08:26.998714 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4w8l\" (UniqueName: \"kubernetes.io/projected/7efebc3a-dab4-4039-8fe1-e2e361666b97-kube-api-access-p4w8l\") pod \"ceilometer-0\" (UID: \"7efebc3a-dab4-4039-8fe1-e2e361666b97\") " pod="openstack/ceilometer-0" Oct 02 10:08:27 crc kubenswrapper[4934]: I1002 10:08:27.032157 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:08:27 crc kubenswrapper[4934]: I1002 10:08:27.468788 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:08:27 crc kubenswrapper[4934]: W1002 10:08:27.474848 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7efebc3a_dab4_4039_8fe1_e2e361666b97.slice/crio-499478fd74d1261b6bdcc1d446e6a53507814eaea857293a590b372dc105f92b WatchSource:0}: Error finding container 499478fd74d1261b6bdcc1d446e6a53507814eaea857293a590b372dc105f92b: Status 404 returned error can't find the container with id 499478fd74d1261b6bdcc1d446e6a53507814eaea857293a590b372dc105f92b Oct 02 10:08:27 crc kubenswrapper[4934]: I1002 10:08:27.599058 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7efebc3a-dab4-4039-8fe1-e2e361666b97","Type":"ContainerStarted","Data":"499478fd74d1261b6bdcc1d446e6a53507814eaea857293a590b372dc105f92b"} Oct 02 10:08:27 crc kubenswrapper[4934]: I1002 10:08:27.655772 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 10:08:27 crc kubenswrapper[4934]: I1002 10:08:27.655827 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 10:08:27 crc kubenswrapper[4934]: I1002 10:08:27.716001 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 10:08:27 crc kubenswrapper[4934]: I1002 10:08:27.716083 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 10:08:28 crc kubenswrapper[4934]: I1002 10:08:28.018230 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kh9px" Oct 02 10:08:28 crc kubenswrapper[4934]: I1002 10:08:28.096109 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvn4d\" (UniqueName: \"kubernetes.io/projected/09b1ea6c-4d4a-427d-a66e-e3db8b198316-kube-api-access-zvn4d\") pod \"09b1ea6c-4d4a-427d-a66e-e3db8b198316\" (UID: \"09b1ea6c-4d4a-427d-a66e-e3db8b198316\") " Oct 02 10:08:28 crc kubenswrapper[4934]: I1002 10:08:28.101382 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09b1ea6c-4d4a-427d-a66e-e3db8b198316-kube-api-access-zvn4d" (OuterVolumeSpecName: "kube-api-access-zvn4d") pod "09b1ea6c-4d4a-427d-a66e-e3db8b198316" (UID: "09b1ea6c-4d4a-427d-a66e-e3db8b198316"). InnerVolumeSpecName "kube-api-access-zvn4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:08:28 crc kubenswrapper[4934]: I1002 10:08:28.163475 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4z9rv" Oct 02 10:08:28 crc kubenswrapper[4934]: I1002 10:08:28.171060 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-h84gj" Oct 02 10:08:28 crc kubenswrapper[4934]: I1002 10:08:28.198099 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvn4d\" (UniqueName: \"kubernetes.io/projected/09b1ea6c-4d4a-427d-a66e-e3db8b198316-kube-api-access-zvn4d\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:28 crc kubenswrapper[4934]: I1002 10:08:28.299268 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qkxx\" (UniqueName: \"kubernetes.io/projected/26f3aea9-8e30-4ec3-b440-c00fe8f624da-kube-api-access-7qkxx\") pod \"26f3aea9-8e30-4ec3-b440-c00fe8f624da\" (UID: \"26f3aea9-8e30-4ec3-b440-c00fe8f624da\") " Oct 02 10:08:28 crc kubenswrapper[4934]: I1002 10:08:28.299508 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dc7px\" (UniqueName: \"kubernetes.io/projected/f57c470b-0a2a-4341-ae16-6d5ac848ade0-kube-api-access-dc7px\") pod \"f57c470b-0a2a-4341-ae16-6d5ac848ade0\" (UID: \"f57c470b-0a2a-4341-ae16-6d5ac848ade0\") " Oct 02 10:08:28 crc kubenswrapper[4934]: I1002 10:08:28.302561 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26f3aea9-8e30-4ec3-b440-c00fe8f624da-kube-api-access-7qkxx" (OuterVolumeSpecName: "kube-api-access-7qkxx") pod "26f3aea9-8e30-4ec3-b440-c00fe8f624da" (UID: "26f3aea9-8e30-4ec3-b440-c00fe8f624da"). InnerVolumeSpecName "kube-api-access-7qkxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:08:28 crc kubenswrapper[4934]: I1002 10:08:28.302922 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f57c470b-0a2a-4341-ae16-6d5ac848ade0-kube-api-access-dc7px" (OuterVolumeSpecName: "kube-api-access-dc7px") pod "f57c470b-0a2a-4341-ae16-6d5ac848ade0" (UID: "f57c470b-0a2a-4341-ae16-6d5ac848ade0"). InnerVolumeSpecName "kube-api-access-dc7px". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:08:28 crc kubenswrapper[4934]: I1002 10:08:28.402029 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dc7px\" (UniqueName: \"kubernetes.io/projected/f57c470b-0a2a-4341-ae16-6d5ac848ade0-kube-api-access-dc7px\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:28 crc kubenswrapper[4934]: I1002 10:08:28.402067 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qkxx\" (UniqueName: \"kubernetes.io/projected/26f3aea9-8e30-4ec3-b440-c00fe8f624da-kube-api-access-7qkxx\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:28 crc kubenswrapper[4934]: I1002 10:08:28.613638 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-kh9px" event={"ID":"09b1ea6c-4d4a-427d-a66e-e3db8b198316","Type":"ContainerDied","Data":"1ae6fabae2fa27df456226420f3311d2a3908de63e7a14c5ee520811a70e98c7"} Oct 02 10:08:28 crc kubenswrapper[4934]: I1002 10:08:28.614024 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ae6fabae2fa27df456226420f3311d2a3908de63e7a14c5ee520811a70e98c7" Oct 02 10:08:28 crc kubenswrapper[4934]: I1002 10:08:28.614109 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kh9px" Oct 02 10:08:28 crc kubenswrapper[4934]: I1002 10:08:28.628443 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-h84gj" event={"ID":"26f3aea9-8e30-4ec3-b440-c00fe8f624da","Type":"ContainerDied","Data":"7d4c5613a1af27fbd439a142db3cfcb83fb8da809b02d31244b45b9418034766"} Oct 02 10:08:28 crc kubenswrapper[4934]: I1002 10:08:28.628497 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-h84gj" Oct 02 10:08:28 crc kubenswrapper[4934]: I1002 10:08:28.628495 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d4c5613a1af27fbd439a142db3cfcb83fb8da809b02d31244b45b9418034766" Oct 02 10:08:28 crc kubenswrapper[4934]: I1002 10:08:28.630318 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7efebc3a-dab4-4039-8fe1-e2e361666b97","Type":"ContainerStarted","Data":"596a9638ff8bc186ea78405fe4a13b4a6fd360b0e7dc381063046280119ce492"} Oct 02 10:08:28 crc kubenswrapper[4934]: I1002 10:08:28.633801 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4z9rv" Oct 02 10:08:28 crc kubenswrapper[4934]: I1002 10:08:28.640740 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4z9rv" event={"ID":"f57c470b-0a2a-4341-ae16-6d5ac848ade0","Type":"ContainerDied","Data":"98c96cadec8fc9f74eb7f134ce267562f51435dd24afc8591c36ed501d992fa8"} Oct 02 10:08:28 crc kubenswrapper[4934]: I1002 10:08:28.640780 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98c96cadec8fc9f74eb7f134ce267562f51435dd24afc8591c36ed501d992fa8" Oct 02 10:08:28 crc kubenswrapper[4934]: I1002 10:08:28.640801 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 10:08:28 crc kubenswrapper[4934]: I1002 10:08:28.640814 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 10:08:29 crc kubenswrapper[4934]: I1002 10:08:29.647090 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7efebc3a-dab4-4039-8fe1-e2e361666b97","Type":"ContainerStarted","Data":"c4b85f44d13837d0d67d362c6c36287790b41f41ca3394c6f3030c5380b61f61"} Oct 02 10:08:30 crc kubenswrapper[4934]: I1002 10:08:30.632334 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 10:08:30 crc kubenswrapper[4934]: I1002 10:08:30.668191 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7efebc3a-dab4-4039-8fe1-e2e361666b97","Type":"ContainerStarted","Data":"da5df09079b1d3f6d4ac17c816ab87d91d6d2c92685faad133b239e783ec1b78"} Oct 02 10:08:30 crc kubenswrapper[4934]: I1002 10:08:30.668210 4934 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 10:08:30 crc kubenswrapper[4934]: I1002 10:08:30.814146 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 10:08:31 crc kubenswrapper[4934]: I1002 10:08:31.929902 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 10:08:31 crc kubenswrapper[4934]: I1002 10:08:31.930141 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 10:08:31 crc kubenswrapper[4934]: I1002 10:08:31.970096 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 10:08:31 crc kubenswrapper[4934]: I1002 10:08:31.980483 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 10:08:32 crc kubenswrapper[4934]: I1002 10:08:32.692300 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7efebc3a-dab4-4039-8fe1-e2e361666b97","Type":"ContainerStarted","Data":"778858204eda2f3afea5a82f201f710b01e4d8cf86073c91550ae6845abcbeec"} Oct 02 10:08:32 crc kubenswrapper[4934]: I1002 10:08:32.692712 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 10:08:32 crc kubenswrapper[4934]: I1002 10:08:32.692732 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 10:08:32 crc kubenswrapper[4934]: I1002 10:08:32.727445 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.628376178 podStartE2EDuration="6.727417056s" podCreationTimestamp="2025-10-02 10:08:26 +0000 UTC" firstStartedPulling="2025-10-02 10:08:27.477207029 +0000 UTC m=+1179.229848551" lastFinishedPulling="2025-10-02 10:08:31.576247907 +0000 UTC m=+1183.328889429" observedRunningTime="2025-10-02 10:08:32.718347336 +0000 UTC m=+1184.470988868" watchObservedRunningTime="2025-10-02 10:08:32.727417056 +0000 UTC m=+1184.480058598" Oct 02 10:08:32 crc kubenswrapper[4934]: I1002 10:08:32.787846 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:08:33 crc kubenswrapper[4934]: I1002 10:08:33.700820 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 10:08:34 crc kubenswrapper[4934]: I1002 10:08:34.665601 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 10:08:34 crc kubenswrapper[4934]: I1002 10:08:34.680894 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 10:08:34 crc kubenswrapper[4934]: I1002 10:08:34.714211 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7efebc3a-dab4-4039-8fe1-e2e361666b97" containerName="ceilometer-central-agent" containerID="cri-o://596a9638ff8bc186ea78405fe4a13b4a6fd360b0e7dc381063046280119ce492" gracePeriod=30 Oct 02 10:08:34 crc kubenswrapper[4934]: I1002 10:08:34.715335 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7efebc3a-dab4-4039-8fe1-e2e361666b97" containerName="proxy-httpd" containerID="cri-o://778858204eda2f3afea5a82f201f710b01e4d8cf86073c91550ae6845abcbeec" gracePeriod=30 Oct 02 10:08:34 crc kubenswrapper[4934]: I1002 10:08:34.715398 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7efebc3a-dab4-4039-8fe1-e2e361666b97" containerName="sg-core" containerID="cri-o://da5df09079b1d3f6d4ac17c816ab87d91d6d2c92685faad133b239e783ec1b78" gracePeriod=30 Oct 02 10:08:34 crc kubenswrapper[4934]: I1002 10:08:34.715446 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7efebc3a-dab4-4039-8fe1-e2e361666b97" containerName="ceilometer-notification-agent" containerID="cri-o://c4b85f44d13837d0d67d362c6c36287790b41f41ca3394c6f3030c5380b61f61" gracePeriod=30 Oct 02 10:08:34 crc kubenswrapper[4934]: I1002 10:08:34.793654 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-26a6-account-create-54z6g"] Oct 02 10:08:34 crc kubenswrapper[4934]: E1002 10:08:34.794150 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26f3aea9-8e30-4ec3-b440-c00fe8f624da" containerName="mariadb-database-create" Oct 02 10:08:34 crc kubenswrapper[4934]: I1002 10:08:34.794175 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="26f3aea9-8e30-4ec3-b440-c00fe8f624da" containerName="mariadb-database-create" Oct 02 10:08:34 crc kubenswrapper[4934]: E1002 10:08:34.794188 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f57c470b-0a2a-4341-ae16-6d5ac848ade0" containerName="mariadb-database-create" Oct 02 10:08:34 crc kubenswrapper[4934]: I1002 10:08:34.794196 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f57c470b-0a2a-4341-ae16-6d5ac848ade0" containerName="mariadb-database-create" Oct 02 10:08:34 crc kubenswrapper[4934]: E1002 10:08:34.794222 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09b1ea6c-4d4a-427d-a66e-e3db8b198316" containerName="mariadb-database-create" Oct 02 10:08:34 crc kubenswrapper[4934]: I1002 10:08:34.794230 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="09b1ea6c-4d4a-427d-a66e-e3db8b198316" containerName="mariadb-database-create" Oct 02 10:08:34 crc kubenswrapper[4934]: I1002 10:08:34.794461 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="09b1ea6c-4d4a-427d-a66e-e3db8b198316" containerName="mariadb-database-create" Oct 02 10:08:34 crc kubenswrapper[4934]: I1002 10:08:34.794489 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f57c470b-0a2a-4341-ae16-6d5ac848ade0" containerName="mariadb-database-create" Oct 02 10:08:34 crc kubenswrapper[4934]: I1002 10:08:34.794500 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="26f3aea9-8e30-4ec3-b440-c00fe8f624da" containerName="mariadb-database-create" Oct 02 10:08:34 crc kubenswrapper[4934]: I1002 10:08:34.795260 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-26a6-account-create-54z6g" Oct 02 10:08:34 crc kubenswrapper[4934]: I1002 10:08:34.802220 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 02 10:08:34 crc kubenswrapper[4934]: I1002 10:08:34.807698 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-26a6-account-create-54z6g"] Oct 02 10:08:34 crc kubenswrapper[4934]: I1002 10:08:34.826218 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzk48\" (UniqueName: \"kubernetes.io/projected/53b4d698-fa73-48ad-99c6-4f0aa06ea4ab-kube-api-access-dzk48\") pod \"nova-api-26a6-account-create-54z6g\" (UID: \"53b4d698-fa73-48ad-99c6-4f0aa06ea4ab\") " pod="openstack/nova-api-26a6-account-create-54z6g" Oct 02 10:08:34 crc kubenswrapper[4934]: I1002 10:08:34.928399 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzk48\" (UniqueName: \"kubernetes.io/projected/53b4d698-fa73-48ad-99c6-4f0aa06ea4ab-kube-api-access-dzk48\") pod \"nova-api-26a6-account-create-54z6g\" (UID: \"53b4d698-fa73-48ad-99c6-4f0aa06ea4ab\") " pod="openstack/nova-api-26a6-account-create-54z6g" Oct 02 10:08:34 crc kubenswrapper[4934]: I1002 10:08:34.950881 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzk48\" (UniqueName: \"kubernetes.io/projected/53b4d698-fa73-48ad-99c6-4f0aa06ea4ab-kube-api-access-dzk48\") pod \"nova-api-26a6-account-create-54z6g\" (UID: \"53b4d698-fa73-48ad-99c6-4f0aa06ea4ab\") " pod="openstack/nova-api-26a6-account-create-54z6g" Oct 02 10:08:34 crc kubenswrapper[4934]: I1002 10:08:34.957711 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-9b8b-account-create-v55kd"] Oct 02 10:08:34 crc kubenswrapper[4934]: I1002 10:08:34.958893 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9b8b-account-create-v55kd" Oct 02 10:08:34 crc kubenswrapper[4934]: I1002 10:08:34.961066 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 02 10:08:34 crc kubenswrapper[4934]: I1002 10:08:34.968921 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-9b8b-account-create-v55kd"] Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.030669 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbwwx\" (UniqueName: \"kubernetes.io/projected/3d01321b-b929-4eb8-a780-7a4b4adae54f-kube-api-access-wbwwx\") pod \"nova-cell0-9b8b-account-create-v55kd\" (UID: \"3d01321b-b929-4eb8-a780-7a4b4adae54f\") " pod="openstack/nova-cell0-9b8b-account-create-v55kd" Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.123121 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-26a6-account-create-54z6g" Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.128687 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-ba62-account-create-76fqw"] Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.132090 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbwwx\" (UniqueName: \"kubernetes.io/projected/3d01321b-b929-4eb8-a780-7a4b4adae54f-kube-api-access-wbwwx\") pod \"nova-cell0-9b8b-account-create-v55kd\" (UID: \"3d01321b-b929-4eb8-a780-7a4b4adae54f\") " pod="openstack/nova-cell0-9b8b-account-create-v55kd" Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.134981 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ba62-account-create-76fqw" Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.140500 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-ba62-account-create-76fqw"] Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.141046 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.153282 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbwwx\" (UniqueName: \"kubernetes.io/projected/3d01321b-b929-4eb8-a780-7a4b4adae54f-kube-api-access-wbwwx\") pod \"nova-cell0-9b8b-account-create-v55kd\" (UID: \"3d01321b-b929-4eb8-a780-7a4b4adae54f\") " pod="openstack/nova-cell0-9b8b-account-create-v55kd" Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.234386 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cgmw\" (UniqueName: \"kubernetes.io/projected/99a3a542-8f10-4571-9424-dee0602b290e-kube-api-access-7cgmw\") pod \"nova-cell1-ba62-account-create-76fqw\" (UID: \"99a3a542-8f10-4571-9424-dee0602b290e\") " pod="openstack/nova-cell1-ba62-account-create-76fqw" Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.329045 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9b8b-account-create-v55kd" Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.336346 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cgmw\" (UniqueName: \"kubernetes.io/projected/99a3a542-8f10-4571-9424-dee0602b290e-kube-api-access-7cgmw\") pod \"nova-cell1-ba62-account-create-76fqw\" (UID: \"99a3a542-8f10-4571-9424-dee0602b290e\") " pod="openstack/nova-cell1-ba62-account-create-76fqw" Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.357317 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cgmw\" (UniqueName: \"kubernetes.io/projected/99a3a542-8f10-4571-9424-dee0602b290e-kube-api-access-7cgmw\") pod \"nova-cell1-ba62-account-create-76fqw\" (UID: \"99a3a542-8f10-4571-9424-dee0602b290e\") " pod="openstack/nova-cell1-ba62-account-create-76fqw" Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.586868 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ba62-account-create-76fqw" Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.603957 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-26a6-account-create-54z6g"] Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.666177 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.728561 4934 generic.go:334] "Generic (PLEG): container finished" podID="7efebc3a-dab4-4039-8fe1-e2e361666b97" containerID="778858204eda2f3afea5a82f201f710b01e4d8cf86073c91550ae6845abcbeec" exitCode=0 Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.728875 4934 generic.go:334] "Generic (PLEG): container finished" podID="7efebc3a-dab4-4039-8fe1-e2e361666b97" containerID="da5df09079b1d3f6d4ac17c816ab87d91d6d2c92685faad133b239e783ec1b78" exitCode=2 Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.728886 4934 generic.go:334] "Generic (PLEG): container finished" podID="7efebc3a-dab4-4039-8fe1-e2e361666b97" containerID="c4b85f44d13837d0d67d362c6c36287790b41f41ca3394c6f3030c5380b61f61" exitCode=0 Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.728896 4934 generic.go:334] "Generic (PLEG): container finished" podID="7efebc3a-dab4-4039-8fe1-e2e361666b97" containerID="596a9638ff8bc186ea78405fe4a13b4a6fd360b0e7dc381063046280119ce492" exitCode=0 Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.728791 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7efebc3a-dab4-4039-8fe1-e2e361666b97","Type":"ContainerDied","Data":"778858204eda2f3afea5a82f201f710b01e4d8cf86073c91550ae6845abcbeec"} Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.728934 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7efebc3a-dab4-4039-8fe1-e2e361666b97","Type":"ContainerDied","Data":"da5df09079b1d3f6d4ac17c816ab87d91d6d2c92685faad133b239e783ec1b78"} Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.728886 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.728956 4934 scope.go:117] "RemoveContainer" containerID="778858204eda2f3afea5a82f201f710b01e4d8cf86073c91550ae6845abcbeec" Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.728946 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7efebc3a-dab4-4039-8fe1-e2e361666b97","Type":"ContainerDied","Data":"c4b85f44d13837d0d67d362c6c36287790b41f41ca3394c6f3030c5380b61f61"} Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.729150 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7efebc3a-dab4-4039-8fe1-e2e361666b97","Type":"ContainerDied","Data":"596a9638ff8bc186ea78405fe4a13b4a6fd360b0e7dc381063046280119ce492"} Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.729160 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7efebc3a-dab4-4039-8fe1-e2e361666b97","Type":"ContainerDied","Data":"499478fd74d1261b6bdcc1d446e6a53507814eaea857293a590b372dc105f92b"} Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.734194 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-26a6-account-create-54z6g" event={"ID":"53b4d698-fa73-48ad-99c6-4f0aa06ea4ab","Type":"ContainerStarted","Data":"bd50d4713fecead52ff2fb2a387c5ff6bc517408de733c686e7e37e530c63d2f"} Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.765201 4934 scope.go:117] "RemoveContainer" containerID="da5df09079b1d3f6d4ac17c816ab87d91d6d2c92685faad133b239e783ec1b78" Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.833199 4934 scope.go:117] "RemoveContainer" containerID="c4b85f44d13837d0d67d362c6c36287790b41f41ca3394c6f3030c5380b61f61" Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.843100 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-9b8b-account-create-v55kd"] Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.848465 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7efebc3a-dab4-4039-8fe1-e2e361666b97-log-httpd\") pod \"7efebc3a-dab4-4039-8fe1-e2e361666b97\" (UID: \"7efebc3a-dab4-4039-8fe1-e2e361666b97\") " Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.848515 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7efebc3a-dab4-4039-8fe1-e2e361666b97-config-data\") pod \"7efebc3a-dab4-4039-8fe1-e2e361666b97\" (UID: \"7efebc3a-dab4-4039-8fe1-e2e361666b97\") " Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.848544 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7efebc3a-dab4-4039-8fe1-e2e361666b97-combined-ca-bundle\") pod \"7efebc3a-dab4-4039-8fe1-e2e361666b97\" (UID: \"7efebc3a-dab4-4039-8fe1-e2e361666b97\") " Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.848572 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7efebc3a-dab4-4039-8fe1-e2e361666b97-sg-core-conf-yaml\") pod \"7efebc3a-dab4-4039-8fe1-e2e361666b97\" (UID: \"7efebc3a-dab4-4039-8fe1-e2e361666b97\") " Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.848621 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7efebc3a-dab4-4039-8fe1-e2e361666b97-run-httpd\") pod \"7efebc3a-dab4-4039-8fe1-e2e361666b97\" (UID: \"7efebc3a-dab4-4039-8fe1-e2e361666b97\") " Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.848679 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7efebc3a-dab4-4039-8fe1-e2e361666b97-scripts\") pod \"7efebc3a-dab4-4039-8fe1-e2e361666b97\" (UID: \"7efebc3a-dab4-4039-8fe1-e2e361666b97\") " Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.848738 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4w8l\" (UniqueName: \"kubernetes.io/projected/7efebc3a-dab4-4039-8fe1-e2e361666b97-kube-api-access-p4w8l\") pod \"7efebc3a-dab4-4039-8fe1-e2e361666b97\" (UID: \"7efebc3a-dab4-4039-8fe1-e2e361666b97\") " Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.852831 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7efebc3a-dab4-4039-8fe1-e2e361666b97-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7efebc3a-dab4-4039-8fe1-e2e361666b97" (UID: "7efebc3a-dab4-4039-8fe1-e2e361666b97"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.853358 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7efebc3a-dab4-4039-8fe1-e2e361666b97-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7efebc3a-dab4-4039-8fe1-e2e361666b97" (UID: "7efebc3a-dab4-4039-8fe1-e2e361666b97"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.870428 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7efebc3a-dab4-4039-8fe1-e2e361666b97-kube-api-access-p4w8l" (OuterVolumeSpecName: "kube-api-access-p4w8l") pod "7efebc3a-dab4-4039-8fe1-e2e361666b97" (UID: "7efebc3a-dab4-4039-8fe1-e2e361666b97"). InnerVolumeSpecName "kube-api-access-p4w8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.882730 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7efebc3a-dab4-4039-8fe1-e2e361666b97-scripts" (OuterVolumeSpecName: "scripts") pod "7efebc3a-dab4-4039-8fe1-e2e361666b97" (UID: "7efebc3a-dab4-4039-8fe1-e2e361666b97"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.890843 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7efebc3a-dab4-4039-8fe1-e2e361666b97-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7efebc3a-dab4-4039-8fe1-e2e361666b97" (UID: "7efebc3a-dab4-4039-8fe1-e2e361666b97"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.942287 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7efebc3a-dab4-4039-8fe1-e2e361666b97-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7efebc3a-dab4-4039-8fe1-e2e361666b97" (UID: "7efebc3a-dab4-4039-8fe1-e2e361666b97"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.953254 4934 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7efebc3a-dab4-4039-8fe1-e2e361666b97-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.953288 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7efebc3a-dab4-4039-8fe1-e2e361666b97-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.953300 4934 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7efebc3a-dab4-4039-8fe1-e2e361666b97-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.953308 4934 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7efebc3a-dab4-4039-8fe1-e2e361666b97-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.953317 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7efebc3a-dab4-4039-8fe1-e2e361666b97-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.953325 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4w8l\" (UniqueName: \"kubernetes.io/projected/7efebc3a-dab4-4039-8fe1-e2e361666b97-kube-api-access-p4w8l\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:35 crc kubenswrapper[4934]: I1002 10:08:35.968129 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7efebc3a-dab4-4039-8fe1-e2e361666b97-config-data" (OuterVolumeSpecName: "config-data") pod "7efebc3a-dab4-4039-8fe1-e2e361666b97" (UID: "7efebc3a-dab4-4039-8fe1-e2e361666b97"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.055738 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7efebc3a-dab4-4039-8fe1-e2e361666b97-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.072099 4934 scope.go:117] "RemoveContainer" containerID="596a9638ff8bc186ea78405fe4a13b4a6fd360b0e7dc381063046280119ce492" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.074155 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-ba62-account-create-76fqw"] Oct 02 10:08:36 crc kubenswrapper[4934]: W1002 10:08:36.090953 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99a3a542_8f10_4571_9424_dee0602b290e.slice/crio-1e97374fe08079f5328104f29dd6d86c520ae2a3060a6df282707d440f13c730 WatchSource:0}: Error finding container 1e97374fe08079f5328104f29dd6d86c520ae2a3060a6df282707d440f13c730: Status 404 returned error can't find the container with id 1e97374fe08079f5328104f29dd6d86c520ae2a3060a6df282707d440f13c730 Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.107608 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.124657 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.134322 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:08:36 crc kubenswrapper[4934]: E1002 10:08:36.134833 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7efebc3a-dab4-4039-8fe1-e2e361666b97" containerName="ceilometer-notification-agent" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.134857 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="7efebc3a-dab4-4039-8fe1-e2e361666b97" containerName="ceilometer-notification-agent" Oct 02 10:08:36 crc kubenswrapper[4934]: E1002 10:08:36.134941 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7efebc3a-dab4-4039-8fe1-e2e361666b97" containerName="sg-core" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.134950 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="7efebc3a-dab4-4039-8fe1-e2e361666b97" containerName="sg-core" Oct 02 10:08:36 crc kubenswrapper[4934]: E1002 10:08:36.134963 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7efebc3a-dab4-4039-8fe1-e2e361666b97" containerName="proxy-httpd" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.134969 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="7efebc3a-dab4-4039-8fe1-e2e361666b97" containerName="proxy-httpd" Oct 02 10:08:36 crc kubenswrapper[4934]: E1002 10:08:36.135023 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7efebc3a-dab4-4039-8fe1-e2e361666b97" containerName="ceilometer-central-agent" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.135034 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="7efebc3a-dab4-4039-8fe1-e2e361666b97" containerName="ceilometer-central-agent" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.135368 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="7efebc3a-dab4-4039-8fe1-e2e361666b97" containerName="sg-core" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.135384 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="7efebc3a-dab4-4039-8fe1-e2e361666b97" containerName="proxy-httpd" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.135427 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="7efebc3a-dab4-4039-8fe1-e2e361666b97" containerName="ceilometer-central-agent" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.135441 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="7efebc3a-dab4-4039-8fe1-e2e361666b97" containerName="ceilometer-notification-agent" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.137442 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.143969 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.144229 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.153312 4934 scope.go:117] "RemoveContainer" containerID="778858204eda2f3afea5a82f201f710b01e4d8cf86073c91550ae6845abcbeec" Oct 02 10:08:36 crc kubenswrapper[4934]: E1002 10:08:36.154666 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"778858204eda2f3afea5a82f201f710b01e4d8cf86073c91550ae6845abcbeec\": container with ID starting with 778858204eda2f3afea5a82f201f710b01e4d8cf86073c91550ae6845abcbeec not found: ID does not exist" containerID="778858204eda2f3afea5a82f201f710b01e4d8cf86073c91550ae6845abcbeec" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.154711 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"778858204eda2f3afea5a82f201f710b01e4d8cf86073c91550ae6845abcbeec"} err="failed to get container status \"778858204eda2f3afea5a82f201f710b01e4d8cf86073c91550ae6845abcbeec\": rpc error: code = NotFound desc = could not find container \"778858204eda2f3afea5a82f201f710b01e4d8cf86073c91550ae6845abcbeec\": container with ID starting with 778858204eda2f3afea5a82f201f710b01e4d8cf86073c91550ae6845abcbeec not found: ID does not exist" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.154739 4934 scope.go:117] "RemoveContainer" containerID="da5df09079b1d3f6d4ac17c816ab87d91d6d2c92685faad133b239e783ec1b78" Oct 02 10:08:36 crc kubenswrapper[4934]: E1002 10:08:36.155835 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da5df09079b1d3f6d4ac17c816ab87d91d6d2c92685faad133b239e783ec1b78\": container with ID starting with da5df09079b1d3f6d4ac17c816ab87d91d6d2c92685faad133b239e783ec1b78 not found: ID does not exist" containerID="da5df09079b1d3f6d4ac17c816ab87d91d6d2c92685faad133b239e783ec1b78" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.155880 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da5df09079b1d3f6d4ac17c816ab87d91d6d2c92685faad133b239e783ec1b78"} err="failed to get container status \"da5df09079b1d3f6d4ac17c816ab87d91d6d2c92685faad133b239e783ec1b78\": rpc error: code = NotFound desc = could not find container \"da5df09079b1d3f6d4ac17c816ab87d91d6d2c92685faad133b239e783ec1b78\": container with ID starting with da5df09079b1d3f6d4ac17c816ab87d91d6d2c92685faad133b239e783ec1b78 not found: ID does not exist" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.155911 4934 scope.go:117] "RemoveContainer" containerID="c4b85f44d13837d0d67d362c6c36287790b41f41ca3394c6f3030c5380b61f61" Oct 02 10:08:36 crc kubenswrapper[4934]: E1002 10:08:36.156725 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4b85f44d13837d0d67d362c6c36287790b41f41ca3394c6f3030c5380b61f61\": container with ID starting with c4b85f44d13837d0d67d362c6c36287790b41f41ca3394c6f3030c5380b61f61 not found: ID does not exist" containerID="c4b85f44d13837d0d67d362c6c36287790b41f41ca3394c6f3030c5380b61f61" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.156754 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4b85f44d13837d0d67d362c6c36287790b41f41ca3394c6f3030c5380b61f61"} err="failed to get container status \"c4b85f44d13837d0d67d362c6c36287790b41f41ca3394c6f3030c5380b61f61\": rpc error: code = NotFound desc = could not find container \"c4b85f44d13837d0d67d362c6c36287790b41f41ca3394c6f3030c5380b61f61\": container with ID starting with c4b85f44d13837d0d67d362c6c36287790b41f41ca3394c6f3030c5380b61f61 not found: ID does not exist" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.156768 4934 scope.go:117] "RemoveContainer" containerID="596a9638ff8bc186ea78405fe4a13b4a6fd360b0e7dc381063046280119ce492" Oct 02 10:08:36 crc kubenswrapper[4934]: E1002 10:08:36.157478 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"596a9638ff8bc186ea78405fe4a13b4a6fd360b0e7dc381063046280119ce492\": container with ID starting with 596a9638ff8bc186ea78405fe4a13b4a6fd360b0e7dc381063046280119ce492 not found: ID does not exist" containerID="596a9638ff8bc186ea78405fe4a13b4a6fd360b0e7dc381063046280119ce492" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.157500 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"596a9638ff8bc186ea78405fe4a13b4a6fd360b0e7dc381063046280119ce492"} err="failed to get container status \"596a9638ff8bc186ea78405fe4a13b4a6fd360b0e7dc381063046280119ce492\": rpc error: code = NotFound desc = could not find container \"596a9638ff8bc186ea78405fe4a13b4a6fd360b0e7dc381063046280119ce492\": container with ID starting with 596a9638ff8bc186ea78405fe4a13b4a6fd360b0e7dc381063046280119ce492 not found: ID does not exist" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.157513 4934 scope.go:117] "RemoveContainer" containerID="778858204eda2f3afea5a82f201f710b01e4d8cf86073c91550ae6845abcbeec" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.158628 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"778858204eda2f3afea5a82f201f710b01e4d8cf86073c91550ae6845abcbeec"} err="failed to get container status \"778858204eda2f3afea5a82f201f710b01e4d8cf86073c91550ae6845abcbeec\": rpc error: code = NotFound desc = could not find container \"778858204eda2f3afea5a82f201f710b01e4d8cf86073c91550ae6845abcbeec\": container with ID starting with 778858204eda2f3afea5a82f201f710b01e4d8cf86073c91550ae6845abcbeec not found: ID does not exist" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.158647 4934 scope.go:117] "RemoveContainer" containerID="da5df09079b1d3f6d4ac17c816ab87d91d6d2c92685faad133b239e783ec1b78" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.158965 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da5df09079b1d3f6d4ac17c816ab87d91d6d2c92685faad133b239e783ec1b78"} err="failed to get container status \"da5df09079b1d3f6d4ac17c816ab87d91d6d2c92685faad133b239e783ec1b78\": rpc error: code = NotFound desc = could not find container \"da5df09079b1d3f6d4ac17c816ab87d91d6d2c92685faad133b239e783ec1b78\": container with ID starting with da5df09079b1d3f6d4ac17c816ab87d91d6d2c92685faad133b239e783ec1b78 not found: ID does not exist" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.159007 4934 scope.go:117] "RemoveContainer" containerID="c4b85f44d13837d0d67d362c6c36287790b41f41ca3394c6f3030c5380b61f61" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.159299 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4b85f44d13837d0d67d362c6c36287790b41f41ca3394c6f3030c5380b61f61"} err="failed to get container status \"c4b85f44d13837d0d67d362c6c36287790b41f41ca3394c6f3030c5380b61f61\": rpc error: code = NotFound desc = could not find container \"c4b85f44d13837d0d67d362c6c36287790b41f41ca3394c6f3030c5380b61f61\": container with ID starting with c4b85f44d13837d0d67d362c6c36287790b41f41ca3394c6f3030c5380b61f61 not found: ID does not exist" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.159319 4934 scope.go:117] "RemoveContainer" containerID="596a9638ff8bc186ea78405fe4a13b4a6fd360b0e7dc381063046280119ce492" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.159540 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"596a9638ff8bc186ea78405fe4a13b4a6fd360b0e7dc381063046280119ce492"} err="failed to get container status \"596a9638ff8bc186ea78405fe4a13b4a6fd360b0e7dc381063046280119ce492\": rpc error: code = NotFound desc = could not find container \"596a9638ff8bc186ea78405fe4a13b4a6fd360b0e7dc381063046280119ce492\": container with ID starting with 596a9638ff8bc186ea78405fe4a13b4a6fd360b0e7dc381063046280119ce492 not found: ID does not exist" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.159557 4934 scope.go:117] "RemoveContainer" containerID="778858204eda2f3afea5a82f201f710b01e4d8cf86073c91550ae6845abcbeec" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.159860 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"778858204eda2f3afea5a82f201f710b01e4d8cf86073c91550ae6845abcbeec"} err="failed to get container status \"778858204eda2f3afea5a82f201f710b01e4d8cf86073c91550ae6845abcbeec\": rpc error: code = NotFound desc = could not find container \"778858204eda2f3afea5a82f201f710b01e4d8cf86073c91550ae6845abcbeec\": container with ID starting with 778858204eda2f3afea5a82f201f710b01e4d8cf86073c91550ae6845abcbeec not found: ID does not exist" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.159880 4934 scope.go:117] "RemoveContainer" containerID="da5df09079b1d3f6d4ac17c816ab87d91d6d2c92685faad133b239e783ec1b78" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.160325 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.160495 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da5df09079b1d3f6d4ac17c816ab87d91d6d2c92685faad133b239e783ec1b78"} err="failed to get container status \"da5df09079b1d3f6d4ac17c816ab87d91d6d2c92685faad133b239e783ec1b78\": rpc error: code = NotFound desc = could not find container \"da5df09079b1d3f6d4ac17c816ab87d91d6d2c92685faad133b239e783ec1b78\": container with ID starting with da5df09079b1d3f6d4ac17c816ab87d91d6d2c92685faad133b239e783ec1b78 not found: ID does not exist" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.160543 4934 scope.go:117] "RemoveContainer" containerID="c4b85f44d13837d0d67d362c6c36287790b41f41ca3394c6f3030c5380b61f61" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.164262 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4b85f44d13837d0d67d362c6c36287790b41f41ca3394c6f3030c5380b61f61"} err="failed to get container status \"c4b85f44d13837d0d67d362c6c36287790b41f41ca3394c6f3030c5380b61f61\": rpc error: code = NotFound desc = could not find container \"c4b85f44d13837d0d67d362c6c36287790b41f41ca3394c6f3030c5380b61f61\": container with ID starting with c4b85f44d13837d0d67d362c6c36287790b41f41ca3394c6f3030c5380b61f61 not found: ID does not exist" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.164306 4934 scope.go:117] "RemoveContainer" containerID="596a9638ff8bc186ea78405fe4a13b4a6fd360b0e7dc381063046280119ce492" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.167752 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"596a9638ff8bc186ea78405fe4a13b4a6fd360b0e7dc381063046280119ce492"} err="failed to get container status \"596a9638ff8bc186ea78405fe4a13b4a6fd360b0e7dc381063046280119ce492\": rpc error: code = NotFound desc = could not find container \"596a9638ff8bc186ea78405fe4a13b4a6fd360b0e7dc381063046280119ce492\": container with ID starting with 596a9638ff8bc186ea78405fe4a13b4a6fd360b0e7dc381063046280119ce492 not found: ID does not exist" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.167799 4934 scope.go:117] "RemoveContainer" containerID="778858204eda2f3afea5a82f201f710b01e4d8cf86073c91550ae6845abcbeec" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.170488 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"778858204eda2f3afea5a82f201f710b01e4d8cf86073c91550ae6845abcbeec"} err="failed to get container status \"778858204eda2f3afea5a82f201f710b01e4d8cf86073c91550ae6845abcbeec\": rpc error: code = NotFound desc = could not find container \"778858204eda2f3afea5a82f201f710b01e4d8cf86073c91550ae6845abcbeec\": container with ID starting with 778858204eda2f3afea5a82f201f710b01e4d8cf86073c91550ae6845abcbeec not found: ID does not exist" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.170545 4934 scope.go:117] "RemoveContainer" containerID="da5df09079b1d3f6d4ac17c816ab87d91d6d2c92685faad133b239e783ec1b78" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.170949 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da5df09079b1d3f6d4ac17c816ab87d91d6d2c92685faad133b239e783ec1b78"} err="failed to get container status \"da5df09079b1d3f6d4ac17c816ab87d91d6d2c92685faad133b239e783ec1b78\": rpc error: code = NotFound desc = could not find container \"da5df09079b1d3f6d4ac17c816ab87d91d6d2c92685faad133b239e783ec1b78\": container with ID starting with da5df09079b1d3f6d4ac17c816ab87d91d6d2c92685faad133b239e783ec1b78 not found: ID does not exist" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.170967 4934 scope.go:117] "RemoveContainer" containerID="c4b85f44d13837d0d67d362c6c36287790b41f41ca3394c6f3030c5380b61f61" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.171295 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4b85f44d13837d0d67d362c6c36287790b41f41ca3394c6f3030c5380b61f61"} err="failed to get container status \"c4b85f44d13837d0d67d362c6c36287790b41f41ca3394c6f3030c5380b61f61\": rpc error: code = NotFound desc = could not find container \"c4b85f44d13837d0d67d362c6c36287790b41f41ca3394c6f3030c5380b61f61\": container with ID starting with c4b85f44d13837d0d67d362c6c36287790b41f41ca3394c6f3030c5380b61f61 not found: ID does not exist" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.171316 4934 scope.go:117] "RemoveContainer" containerID="596a9638ff8bc186ea78405fe4a13b4a6fd360b0e7dc381063046280119ce492" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.171622 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"596a9638ff8bc186ea78405fe4a13b4a6fd360b0e7dc381063046280119ce492"} err="failed to get container status \"596a9638ff8bc186ea78405fe4a13b4a6fd360b0e7dc381063046280119ce492\": rpc error: code = NotFound desc = could not find container \"596a9638ff8bc186ea78405fe4a13b4a6fd360b0e7dc381063046280119ce492\": container with ID starting with 596a9638ff8bc186ea78405fe4a13b4a6fd360b0e7dc381063046280119ce492 not found: ID does not exist" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.259216 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecfa5fcf-afae-4e66-96f3-14b356502921-run-httpd\") pod \"ceilometer-0\" (UID: \"ecfa5fcf-afae-4e66-96f3-14b356502921\") " pod="openstack/ceilometer-0" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.259304 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tgg2\" (UniqueName: \"kubernetes.io/projected/ecfa5fcf-afae-4e66-96f3-14b356502921-kube-api-access-5tgg2\") pod \"ceilometer-0\" (UID: \"ecfa5fcf-afae-4e66-96f3-14b356502921\") " pod="openstack/ceilometer-0" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.259326 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecfa5fcf-afae-4e66-96f3-14b356502921-scripts\") pod \"ceilometer-0\" (UID: \"ecfa5fcf-afae-4e66-96f3-14b356502921\") " pod="openstack/ceilometer-0" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.259425 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecfa5fcf-afae-4e66-96f3-14b356502921-log-httpd\") pod \"ceilometer-0\" (UID: \"ecfa5fcf-afae-4e66-96f3-14b356502921\") " pod="openstack/ceilometer-0" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.259474 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ecfa5fcf-afae-4e66-96f3-14b356502921-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ecfa5fcf-afae-4e66-96f3-14b356502921\") " pod="openstack/ceilometer-0" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.259527 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecfa5fcf-afae-4e66-96f3-14b356502921-config-data\") pod \"ceilometer-0\" (UID: \"ecfa5fcf-afae-4e66-96f3-14b356502921\") " pod="openstack/ceilometer-0" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.259561 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecfa5fcf-afae-4e66-96f3-14b356502921-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ecfa5fcf-afae-4e66-96f3-14b356502921\") " pod="openstack/ceilometer-0" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.362222 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecfa5fcf-afae-4e66-96f3-14b356502921-config-data\") pod \"ceilometer-0\" (UID: \"ecfa5fcf-afae-4e66-96f3-14b356502921\") " pod="openstack/ceilometer-0" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.362622 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecfa5fcf-afae-4e66-96f3-14b356502921-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ecfa5fcf-afae-4e66-96f3-14b356502921\") " pod="openstack/ceilometer-0" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.362692 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecfa5fcf-afae-4e66-96f3-14b356502921-run-httpd\") pod \"ceilometer-0\" (UID: \"ecfa5fcf-afae-4e66-96f3-14b356502921\") " pod="openstack/ceilometer-0" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.362754 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tgg2\" (UniqueName: \"kubernetes.io/projected/ecfa5fcf-afae-4e66-96f3-14b356502921-kube-api-access-5tgg2\") pod \"ceilometer-0\" (UID: \"ecfa5fcf-afae-4e66-96f3-14b356502921\") " pod="openstack/ceilometer-0" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.362771 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecfa5fcf-afae-4e66-96f3-14b356502921-scripts\") pod \"ceilometer-0\" (UID: \"ecfa5fcf-afae-4e66-96f3-14b356502921\") " pod="openstack/ceilometer-0" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.362847 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecfa5fcf-afae-4e66-96f3-14b356502921-log-httpd\") pod \"ceilometer-0\" (UID: \"ecfa5fcf-afae-4e66-96f3-14b356502921\") " pod="openstack/ceilometer-0" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.362888 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ecfa5fcf-afae-4e66-96f3-14b356502921-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ecfa5fcf-afae-4e66-96f3-14b356502921\") " pod="openstack/ceilometer-0" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.363133 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecfa5fcf-afae-4e66-96f3-14b356502921-run-httpd\") pod \"ceilometer-0\" (UID: \"ecfa5fcf-afae-4e66-96f3-14b356502921\") " pod="openstack/ceilometer-0" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.363542 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecfa5fcf-afae-4e66-96f3-14b356502921-log-httpd\") pod \"ceilometer-0\" (UID: \"ecfa5fcf-afae-4e66-96f3-14b356502921\") " pod="openstack/ceilometer-0" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.375811 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecfa5fcf-afae-4e66-96f3-14b356502921-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ecfa5fcf-afae-4e66-96f3-14b356502921\") " pod="openstack/ceilometer-0" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.376457 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecfa5fcf-afae-4e66-96f3-14b356502921-scripts\") pod \"ceilometer-0\" (UID: \"ecfa5fcf-afae-4e66-96f3-14b356502921\") " pod="openstack/ceilometer-0" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.376702 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ecfa5fcf-afae-4e66-96f3-14b356502921-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ecfa5fcf-afae-4e66-96f3-14b356502921\") " pod="openstack/ceilometer-0" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.379175 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecfa5fcf-afae-4e66-96f3-14b356502921-config-data\") pod \"ceilometer-0\" (UID: \"ecfa5fcf-afae-4e66-96f3-14b356502921\") " pod="openstack/ceilometer-0" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.381157 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tgg2\" (UniqueName: \"kubernetes.io/projected/ecfa5fcf-afae-4e66-96f3-14b356502921-kube-api-access-5tgg2\") pod \"ceilometer-0\" (UID: \"ecfa5fcf-afae-4e66-96f3-14b356502921\") " pod="openstack/ceilometer-0" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.495065 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.750105 4934 generic.go:334] "Generic (PLEG): container finished" podID="99a3a542-8f10-4571-9424-dee0602b290e" containerID="da7e910ecc8c66f11827ac4b9c15d0fce9bc3238df05f00389ab9585acf7d276" exitCode=0 Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.750267 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ba62-account-create-76fqw" event={"ID":"99a3a542-8f10-4571-9424-dee0602b290e","Type":"ContainerDied","Data":"da7e910ecc8c66f11827ac4b9c15d0fce9bc3238df05f00389ab9585acf7d276"} Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.750703 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ba62-account-create-76fqw" event={"ID":"99a3a542-8f10-4571-9424-dee0602b290e","Type":"ContainerStarted","Data":"1e97374fe08079f5328104f29dd6d86c520ae2a3060a6df282707d440f13c730"} Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.790412 4934 generic.go:334] "Generic (PLEG): container finished" podID="3d01321b-b929-4eb8-a780-7a4b4adae54f" containerID="3b14288c234464c443d841b2865a70bd53e55ad242d835bc723d8515de567383" exitCode=0 Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.790516 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9b8b-account-create-v55kd" event={"ID":"3d01321b-b929-4eb8-a780-7a4b4adae54f","Type":"ContainerDied","Data":"3b14288c234464c443d841b2865a70bd53e55ad242d835bc723d8515de567383"} Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.790552 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9b8b-account-create-v55kd" event={"ID":"3d01321b-b929-4eb8-a780-7a4b4adae54f","Type":"ContainerStarted","Data":"e3b4c2d2d854583c5d230ae062bfcfd031d9f1e4b7b4507819fe08b0d3940834"} Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.797619 4934 generic.go:334] "Generic (PLEG): container finished" podID="53b4d698-fa73-48ad-99c6-4f0aa06ea4ab" containerID="be7e01c0579a42586b0d2b4e10c3e188d7aa80a13a9f3bf0750dcee6fa06cd0e" exitCode=0 Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.797866 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-26a6-account-create-54z6g" event={"ID":"53b4d698-fa73-48ad-99c6-4f0aa06ea4ab","Type":"ContainerDied","Data":"be7e01c0579a42586b0d2b4e10c3e188d7aa80a13a9f3bf0750dcee6fa06cd0e"} Oct 02 10:08:36 crc kubenswrapper[4934]: I1002 10:08:36.929909 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7efebc3a-dab4-4039-8fe1-e2e361666b97" path="/var/lib/kubelet/pods/7efebc3a-dab4-4039-8fe1-e2e361666b97/volumes" Oct 02 10:08:37 crc kubenswrapper[4934]: I1002 10:08:37.000304 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:08:37 crc kubenswrapper[4934]: I1002 10:08:37.299190 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="edb76b11-9129-495d-b44d-e998c3e8dceb" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.146:3000/\": dial tcp 10.217.0.146:3000: i/o timeout (Client.Timeout exceeded while awaiting headers)" Oct 02 10:08:37 crc kubenswrapper[4934]: I1002 10:08:37.809078 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecfa5fcf-afae-4e66-96f3-14b356502921","Type":"ContainerStarted","Data":"2398e966a7589a94b6b2a97704ab810a68ed18921a76ef565dd9c611f48190de"} Oct 02 10:08:38 crc kubenswrapper[4934]: I1002 10:08:38.246562 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-26a6-account-create-54z6g" Oct 02 10:08:38 crc kubenswrapper[4934]: I1002 10:08:38.353363 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9b8b-account-create-v55kd" Oct 02 10:08:38 crc kubenswrapper[4934]: I1002 10:08:38.355235 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ba62-account-create-76fqw" Oct 02 10:08:38 crc kubenswrapper[4934]: I1002 10:08:38.403683 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzk48\" (UniqueName: \"kubernetes.io/projected/53b4d698-fa73-48ad-99c6-4f0aa06ea4ab-kube-api-access-dzk48\") pod \"53b4d698-fa73-48ad-99c6-4f0aa06ea4ab\" (UID: \"53b4d698-fa73-48ad-99c6-4f0aa06ea4ab\") " Oct 02 10:08:38 crc kubenswrapper[4934]: I1002 10:08:38.411362 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53b4d698-fa73-48ad-99c6-4f0aa06ea4ab-kube-api-access-dzk48" (OuterVolumeSpecName: "kube-api-access-dzk48") pod "53b4d698-fa73-48ad-99c6-4f0aa06ea4ab" (UID: "53b4d698-fa73-48ad-99c6-4f0aa06ea4ab"). InnerVolumeSpecName "kube-api-access-dzk48". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:08:38 crc kubenswrapper[4934]: I1002 10:08:38.504831 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbwwx\" (UniqueName: \"kubernetes.io/projected/3d01321b-b929-4eb8-a780-7a4b4adae54f-kube-api-access-wbwwx\") pod \"3d01321b-b929-4eb8-a780-7a4b4adae54f\" (UID: \"3d01321b-b929-4eb8-a780-7a4b4adae54f\") " Oct 02 10:08:38 crc kubenswrapper[4934]: I1002 10:08:38.505312 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cgmw\" (UniqueName: \"kubernetes.io/projected/99a3a542-8f10-4571-9424-dee0602b290e-kube-api-access-7cgmw\") pod \"99a3a542-8f10-4571-9424-dee0602b290e\" (UID: \"99a3a542-8f10-4571-9424-dee0602b290e\") " Oct 02 10:08:38 crc kubenswrapper[4934]: I1002 10:08:38.505657 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzk48\" (UniqueName: \"kubernetes.io/projected/53b4d698-fa73-48ad-99c6-4f0aa06ea4ab-kube-api-access-dzk48\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:38 crc kubenswrapper[4934]: I1002 10:08:38.507515 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d01321b-b929-4eb8-a780-7a4b4adae54f-kube-api-access-wbwwx" (OuterVolumeSpecName: "kube-api-access-wbwwx") pod "3d01321b-b929-4eb8-a780-7a4b4adae54f" (UID: "3d01321b-b929-4eb8-a780-7a4b4adae54f"). InnerVolumeSpecName "kube-api-access-wbwwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:08:38 crc kubenswrapper[4934]: I1002 10:08:38.509074 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99a3a542-8f10-4571-9424-dee0602b290e-kube-api-access-7cgmw" (OuterVolumeSpecName: "kube-api-access-7cgmw") pod "99a3a542-8f10-4571-9424-dee0602b290e" (UID: "99a3a542-8f10-4571-9424-dee0602b290e"). InnerVolumeSpecName "kube-api-access-7cgmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:08:38 crc kubenswrapper[4934]: I1002 10:08:38.607770 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbwwx\" (UniqueName: \"kubernetes.io/projected/3d01321b-b929-4eb8-a780-7a4b4adae54f-kube-api-access-wbwwx\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:38 crc kubenswrapper[4934]: I1002 10:08:38.607822 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cgmw\" (UniqueName: \"kubernetes.io/projected/99a3a542-8f10-4571-9424-dee0602b290e-kube-api-access-7cgmw\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:38 crc kubenswrapper[4934]: I1002 10:08:38.819078 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9b8b-account-create-v55kd" Oct 02 10:08:38 crc kubenswrapper[4934]: I1002 10:08:38.823675 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9b8b-account-create-v55kd" event={"ID":"3d01321b-b929-4eb8-a780-7a4b4adae54f","Type":"ContainerDied","Data":"e3b4c2d2d854583c5d230ae062bfcfd031d9f1e4b7b4507819fe08b0d3940834"} Oct 02 10:08:38 crc kubenswrapper[4934]: I1002 10:08:38.823710 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3b4c2d2d854583c5d230ae062bfcfd031d9f1e4b7b4507819fe08b0d3940834" Oct 02 10:08:38 crc kubenswrapper[4934]: I1002 10:08:38.826023 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-26a6-account-create-54z6g" event={"ID":"53b4d698-fa73-48ad-99c6-4f0aa06ea4ab","Type":"ContainerDied","Data":"bd50d4713fecead52ff2fb2a387c5ff6bc517408de733c686e7e37e530c63d2f"} Oct 02 10:08:38 crc kubenswrapper[4934]: I1002 10:08:38.826044 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd50d4713fecead52ff2fb2a387c5ff6bc517408de733c686e7e37e530c63d2f" Oct 02 10:08:38 crc kubenswrapper[4934]: I1002 10:08:38.826097 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-26a6-account-create-54z6g" Oct 02 10:08:38 crc kubenswrapper[4934]: I1002 10:08:38.828148 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ba62-account-create-76fqw" event={"ID":"99a3a542-8f10-4571-9424-dee0602b290e","Type":"ContainerDied","Data":"1e97374fe08079f5328104f29dd6d86c520ae2a3060a6df282707d440f13c730"} Oct 02 10:08:38 crc kubenswrapper[4934]: I1002 10:08:38.828173 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e97374fe08079f5328104f29dd6d86c520ae2a3060a6df282707d440f13c730" Oct 02 10:08:38 crc kubenswrapper[4934]: I1002 10:08:38.828231 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ba62-account-create-76fqw" Oct 02 10:08:38 crc kubenswrapper[4934]: I1002 10:08:38.830996 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecfa5fcf-afae-4e66-96f3-14b356502921","Type":"ContainerStarted","Data":"a5362aa5c80c6fcf6a44c0db174e8545d8e68048cb3250a039ff0ee2f1e83c03"} Oct 02 10:08:38 crc kubenswrapper[4934]: I1002 10:08:38.831035 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecfa5fcf-afae-4e66-96f3-14b356502921","Type":"ContainerStarted","Data":"63ed7f16e4336b15b61aea81078bf140c4e03c8913b091e80f8491b0381690a5"} Oct 02 10:08:40 crc kubenswrapper[4934]: I1002 10:08:40.184198 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-vzrk9"] Oct 02 10:08:40 crc kubenswrapper[4934]: E1002 10:08:40.185067 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53b4d698-fa73-48ad-99c6-4f0aa06ea4ab" containerName="mariadb-account-create" Oct 02 10:08:40 crc kubenswrapper[4934]: I1002 10:08:40.185087 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="53b4d698-fa73-48ad-99c6-4f0aa06ea4ab" containerName="mariadb-account-create" Oct 02 10:08:40 crc kubenswrapper[4934]: E1002 10:08:40.185121 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99a3a542-8f10-4571-9424-dee0602b290e" containerName="mariadb-account-create" Oct 02 10:08:40 crc kubenswrapper[4934]: I1002 10:08:40.185130 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="99a3a542-8f10-4571-9424-dee0602b290e" containerName="mariadb-account-create" Oct 02 10:08:40 crc kubenswrapper[4934]: E1002 10:08:40.185147 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d01321b-b929-4eb8-a780-7a4b4adae54f" containerName="mariadb-account-create" Oct 02 10:08:40 crc kubenswrapper[4934]: I1002 10:08:40.185156 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d01321b-b929-4eb8-a780-7a4b4adae54f" containerName="mariadb-account-create" Oct 02 10:08:40 crc kubenswrapper[4934]: I1002 10:08:40.185422 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d01321b-b929-4eb8-a780-7a4b4adae54f" containerName="mariadb-account-create" Oct 02 10:08:40 crc kubenswrapper[4934]: I1002 10:08:40.185460 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="99a3a542-8f10-4571-9424-dee0602b290e" containerName="mariadb-account-create" Oct 02 10:08:40 crc kubenswrapper[4934]: I1002 10:08:40.185473 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="53b4d698-fa73-48ad-99c6-4f0aa06ea4ab" containerName="mariadb-account-create" Oct 02 10:08:40 crc kubenswrapper[4934]: I1002 10:08:40.186298 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-vzrk9" Oct 02 10:08:40 crc kubenswrapper[4934]: I1002 10:08:40.190093 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-96jp7" Oct 02 10:08:40 crc kubenswrapper[4934]: I1002 10:08:40.190242 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 02 10:08:40 crc kubenswrapper[4934]: I1002 10:08:40.190917 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 02 10:08:40 crc kubenswrapper[4934]: I1002 10:08:40.192647 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-vzrk9"] Oct 02 10:08:40 crc kubenswrapper[4934]: I1002 10:08:40.244214 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b-config-data\") pod \"nova-cell0-conductor-db-sync-vzrk9\" (UID: \"bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b\") " pod="openstack/nova-cell0-conductor-db-sync-vzrk9" Oct 02 10:08:40 crc kubenswrapper[4934]: I1002 10:08:40.244603 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-vzrk9\" (UID: \"bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b\") " pod="openstack/nova-cell0-conductor-db-sync-vzrk9" Oct 02 10:08:40 crc kubenswrapper[4934]: I1002 10:08:40.244626 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqzjt\" (UniqueName: \"kubernetes.io/projected/bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b-kube-api-access-hqzjt\") pod \"nova-cell0-conductor-db-sync-vzrk9\" (UID: \"bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b\") " pod="openstack/nova-cell0-conductor-db-sync-vzrk9" Oct 02 10:08:40 crc kubenswrapper[4934]: I1002 10:08:40.244718 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b-scripts\") pod \"nova-cell0-conductor-db-sync-vzrk9\" (UID: \"bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b\") " pod="openstack/nova-cell0-conductor-db-sync-vzrk9" Oct 02 10:08:40 crc kubenswrapper[4934]: I1002 10:08:40.308982 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-856f7d44fb-8lgx8" Oct 02 10:08:40 crc kubenswrapper[4934]: I1002 10:08:40.348829 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-vzrk9\" (UID: \"bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b\") " pod="openstack/nova-cell0-conductor-db-sync-vzrk9" Oct 02 10:08:40 crc kubenswrapper[4934]: I1002 10:08:40.348888 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqzjt\" (UniqueName: \"kubernetes.io/projected/bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b-kube-api-access-hqzjt\") pod \"nova-cell0-conductor-db-sync-vzrk9\" (UID: \"bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b\") " pod="openstack/nova-cell0-conductor-db-sync-vzrk9" Oct 02 10:08:40 crc kubenswrapper[4934]: I1002 10:08:40.348990 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b-scripts\") pod \"nova-cell0-conductor-db-sync-vzrk9\" (UID: \"bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b\") " pod="openstack/nova-cell0-conductor-db-sync-vzrk9" Oct 02 10:08:40 crc kubenswrapper[4934]: I1002 10:08:40.349119 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b-config-data\") pod \"nova-cell0-conductor-db-sync-vzrk9\" (UID: \"bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b\") " pod="openstack/nova-cell0-conductor-db-sync-vzrk9" Oct 02 10:08:40 crc kubenswrapper[4934]: I1002 10:08:40.357918 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-vzrk9\" (UID: \"bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b\") " pod="openstack/nova-cell0-conductor-db-sync-vzrk9" Oct 02 10:08:40 crc kubenswrapper[4934]: I1002 10:08:40.358148 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b-config-data\") pod \"nova-cell0-conductor-db-sync-vzrk9\" (UID: \"bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b\") " pod="openstack/nova-cell0-conductor-db-sync-vzrk9" Oct 02 10:08:40 crc kubenswrapper[4934]: I1002 10:08:40.358401 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b-scripts\") pod \"nova-cell0-conductor-db-sync-vzrk9\" (UID: \"bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b\") " pod="openstack/nova-cell0-conductor-db-sync-vzrk9" Oct 02 10:08:40 crc kubenswrapper[4934]: I1002 10:08:40.376003 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqzjt\" (UniqueName: \"kubernetes.io/projected/bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b-kube-api-access-hqzjt\") pod \"nova-cell0-conductor-db-sync-vzrk9\" (UID: \"bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b\") " pod="openstack/nova-cell0-conductor-db-sync-vzrk9" Oct 02 10:08:40 crc kubenswrapper[4934]: I1002 10:08:40.510629 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-vzrk9" Oct 02 10:08:40 crc kubenswrapper[4934]: I1002 10:08:40.849172 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecfa5fcf-afae-4e66-96f3-14b356502921","Type":"ContainerStarted","Data":"b0a24151d342c969bcd0ce412285ed2ab87f2d65396f71b93b3f1d96e5f90ef3"} Oct 02 10:08:41 crc kubenswrapper[4934]: I1002 10:08:41.229169 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-vzrk9"] Oct 02 10:08:41 crc kubenswrapper[4934]: W1002 10:08:41.230280 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb82cdce_d0b5_48f8_bd1b_d3467fa5a84b.slice/crio-3234fd55176ffcc8aa8a597cff9b3be1ccb0d9f2c86c1c851f4c567c59526647 WatchSource:0}: Error finding container 3234fd55176ffcc8aa8a597cff9b3be1ccb0d9f2c86c1c851f4c567c59526647: Status 404 returned error can't find the container with id 3234fd55176ffcc8aa8a597cff9b3be1ccb0d9f2c86c1c851f4c567c59526647 Oct 02 10:08:41 crc kubenswrapper[4934]: I1002 10:08:41.862294 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-vzrk9" event={"ID":"bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b","Type":"ContainerStarted","Data":"3234fd55176ffcc8aa8a597cff9b3be1ccb0d9f2c86c1c851f4c567c59526647"} Oct 02 10:08:42 crc kubenswrapper[4934]: I1002 10:08:42.874901 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecfa5fcf-afae-4e66-96f3-14b356502921","Type":"ContainerStarted","Data":"6c3ed32906229b970e1b9aac959c6e9222e93273bec8a473c327763dcbfd2d93"} Oct 02 10:08:42 crc kubenswrapper[4934]: I1002 10:08:42.875384 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 10:08:42 crc kubenswrapper[4934]: I1002 10:08:42.908784 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.768938742 podStartE2EDuration="6.908762782s" podCreationTimestamp="2025-10-02 10:08:36 +0000 UTC" firstStartedPulling="2025-10-02 10:08:36.987329715 +0000 UTC m=+1188.739971237" lastFinishedPulling="2025-10-02 10:08:42.127153735 +0000 UTC m=+1193.879795277" observedRunningTime="2025-10-02 10:08:42.899285891 +0000 UTC m=+1194.651927413" watchObservedRunningTime="2025-10-02 10:08:42.908762782 +0000 UTC m=+1194.661404304" Oct 02 10:08:46 crc kubenswrapper[4934]: I1002 10:08:46.432868 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-57456ffd97-rqc8b" Oct 02 10:08:46 crc kubenswrapper[4934]: I1002 10:08:46.495080 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-856f7d44fb-8lgx8"] Oct 02 10:08:46 crc kubenswrapper[4934]: I1002 10:08:46.495534 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-856f7d44fb-8lgx8" podUID="9d8702c0-4b10-41e5-b170-64a356553760" containerName="neutron-api" containerID="cri-o://7b1b88de67d53c8cce24ae235bd6d265be9585b8b71ccf82999a91c8aecefc42" gracePeriod=30 Oct 02 10:08:46 crc kubenswrapper[4934]: I1002 10:08:46.495922 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-856f7d44fb-8lgx8" podUID="9d8702c0-4b10-41e5-b170-64a356553760" containerName="neutron-httpd" containerID="cri-o://a2462b4c5120af0b9bbf15be9bd9b839ac5c8d932d347d54f9c06988a93fd7dd" gracePeriod=30 Oct 02 10:08:46 crc kubenswrapper[4934]: I1002 10:08:46.935108 4934 generic.go:334] "Generic (PLEG): container finished" podID="9d8702c0-4b10-41e5-b170-64a356553760" containerID="a2462b4c5120af0b9bbf15be9bd9b839ac5c8d932d347d54f9c06988a93fd7dd" exitCode=0 Oct 02 10:08:46 crc kubenswrapper[4934]: I1002 10:08:46.943895 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-856f7d44fb-8lgx8" event={"ID":"9d8702c0-4b10-41e5-b170-64a356553760","Type":"ContainerDied","Data":"a2462b4c5120af0b9bbf15be9bd9b839ac5c8d932d347d54f9c06988a93fd7dd"} Oct 02 10:08:48 crc kubenswrapper[4934]: I1002 10:08:48.961890 4934 generic.go:334] "Generic (PLEG): container finished" podID="9d8702c0-4b10-41e5-b170-64a356553760" containerID="7b1b88de67d53c8cce24ae235bd6d265be9585b8b71ccf82999a91c8aecefc42" exitCode=0 Oct 02 10:08:48 crc kubenswrapper[4934]: I1002 10:08:48.962020 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-856f7d44fb-8lgx8" event={"ID":"9d8702c0-4b10-41e5-b170-64a356553760","Type":"ContainerDied","Data":"7b1b88de67d53c8cce24ae235bd6d265be9585b8b71ccf82999a91c8aecefc42"} Oct 02 10:08:49 crc kubenswrapper[4934]: I1002 10:08:49.663601 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:08:49 crc kubenswrapper[4934]: I1002 10:08:49.664348 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ecfa5fcf-afae-4e66-96f3-14b356502921" containerName="ceilometer-central-agent" containerID="cri-o://63ed7f16e4336b15b61aea81078bf140c4e03c8913b091e80f8491b0381690a5" gracePeriod=30 Oct 02 10:08:49 crc kubenswrapper[4934]: I1002 10:08:49.664797 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ecfa5fcf-afae-4e66-96f3-14b356502921" containerName="proxy-httpd" containerID="cri-o://6c3ed32906229b970e1b9aac959c6e9222e93273bec8a473c327763dcbfd2d93" gracePeriod=30 Oct 02 10:08:49 crc kubenswrapper[4934]: I1002 10:08:49.664924 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ecfa5fcf-afae-4e66-96f3-14b356502921" containerName="ceilometer-notification-agent" containerID="cri-o://a5362aa5c80c6fcf6a44c0db174e8545d8e68048cb3250a039ff0ee2f1e83c03" gracePeriod=30 Oct 02 10:08:49 crc kubenswrapper[4934]: I1002 10:08:49.664989 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ecfa5fcf-afae-4e66-96f3-14b356502921" containerName="sg-core" containerID="cri-o://b0a24151d342c969bcd0ce412285ed2ab87f2d65396f71b93b3f1d96e5f90ef3" gracePeriod=30 Oct 02 10:08:50 crc kubenswrapper[4934]: I1002 10:08:50.994868 4934 generic.go:334] "Generic (PLEG): container finished" podID="ecfa5fcf-afae-4e66-96f3-14b356502921" containerID="6c3ed32906229b970e1b9aac959c6e9222e93273bec8a473c327763dcbfd2d93" exitCode=0 Oct 02 10:08:50 crc kubenswrapper[4934]: I1002 10:08:50.994902 4934 generic.go:334] "Generic (PLEG): container finished" podID="ecfa5fcf-afae-4e66-96f3-14b356502921" containerID="b0a24151d342c969bcd0ce412285ed2ab87f2d65396f71b93b3f1d96e5f90ef3" exitCode=2 Oct 02 10:08:50 crc kubenswrapper[4934]: I1002 10:08:50.994910 4934 generic.go:334] "Generic (PLEG): container finished" podID="ecfa5fcf-afae-4e66-96f3-14b356502921" containerID="a5362aa5c80c6fcf6a44c0db174e8545d8e68048cb3250a039ff0ee2f1e83c03" exitCode=0 Oct 02 10:08:50 crc kubenswrapper[4934]: I1002 10:08:50.994919 4934 generic.go:334] "Generic (PLEG): container finished" podID="ecfa5fcf-afae-4e66-96f3-14b356502921" containerID="63ed7f16e4336b15b61aea81078bf140c4e03c8913b091e80f8491b0381690a5" exitCode=0 Oct 02 10:08:50 crc kubenswrapper[4934]: I1002 10:08:50.994945 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecfa5fcf-afae-4e66-96f3-14b356502921","Type":"ContainerDied","Data":"6c3ed32906229b970e1b9aac959c6e9222e93273bec8a473c327763dcbfd2d93"} Oct 02 10:08:50 crc kubenswrapper[4934]: I1002 10:08:50.994989 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecfa5fcf-afae-4e66-96f3-14b356502921","Type":"ContainerDied","Data":"b0a24151d342c969bcd0ce412285ed2ab87f2d65396f71b93b3f1d96e5f90ef3"} Oct 02 10:08:50 crc kubenswrapper[4934]: I1002 10:08:50.995002 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecfa5fcf-afae-4e66-96f3-14b356502921","Type":"ContainerDied","Data":"a5362aa5c80c6fcf6a44c0db174e8545d8e68048cb3250a039ff0ee2f1e83c03"} Oct 02 10:08:50 crc kubenswrapper[4934]: I1002 10:08:50.995011 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecfa5fcf-afae-4e66-96f3-14b356502921","Type":"ContainerDied","Data":"63ed7f16e4336b15b61aea81078bf140c4e03c8913b091e80f8491b0381690a5"} Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.148764 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.207505 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-856f7d44fb-8lgx8" Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.277540 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecfa5fcf-afae-4e66-96f3-14b356502921-config-data\") pod \"ecfa5fcf-afae-4e66-96f3-14b356502921\" (UID: \"ecfa5fcf-afae-4e66-96f3-14b356502921\") " Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.277907 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdbmg\" (UniqueName: \"kubernetes.io/projected/9d8702c0-4b10-41e5-b170-64a356553760-kube-api-access-cdbmg\") pod \"9d8702c0-4b10-41e5-b170-64a356553760\" (UID: \"9d8702c0-4b10-41e5-b170-64a356553760\") " Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.278060 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecfa5fcf-afae-4e66-96f3-14b356502921-log-httpd\") pod \"ecfa5fcf-afae-4e66-96f3-14b356502921\" (UID: \"ecfa5fcf-afae-4e66-96f3-14b356502921\") " Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.278249 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecfa5fcf-afae-4e66-96f3-14b356502921-scripts\") pod \"ecfa5fcf-afae-4e66-96f3-14b356502921\" (UID: \"ecfa5fcf-afae-4e66-96f3-14b356502921\") " Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.278810 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecfa5fcf-afae-4e66-96f3-14b356502921-combined-ca-bundle\") pod \"ecfa5fcf-afae-4e66-96f3-14b356502921\" (UID: \"ecfa5fcf-afae-4e66-96f3-14b356502921\") " Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.278971 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecfa5fcf-afae-4e66-96f3-14b356502921-run-httpd\") pod \"ecfa5fcf-afae-4e66-96f3-14b356502921\" (UID: \"ecfa5fcf-afae-4e66-96f3-14b356502921\") " Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.278656 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecfa5fcf-afae-4e66-96f3-14b356502921-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ecfa5fcf-afae-4e66-96f3-14b356502921" (UID: "ecfa5fcf-afae-4e66-96f3-14b356502921"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.279140 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9d8702c0-4b10-41e5-b170-64a356553760-httpd-config\") pod \"9d8702c0-4b10-41e5-b170-64a356553760\" (UID: \"9d8702c0-4b10-41e5-b170-64a356553760\") " Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.279228 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9d8702c0-4b10-41e5-b170-64a356553760-config\") pod \"9d8702c0-4b10-41e5-b170-64a356553760\" (UID: \"9d8702c0-4b10-41e5-b170-64a356553760\") " Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.279313 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecfa5fcf-afae-4e66-96f3-14b356502921-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ecfa5fcf-afae-4e66-96f3-14b356502921" (UID: "ecfa5fcf-afae-4e66-96f3-14b356502921"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.279337 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ecfa5fcf-afae-4e66-96f3-14b356502921-sg-core-conf-yaml\") pod \"ecfa5fcf-afae-4e66-96f3-14b356502921\" (UID: \"ecfa5fcf-afae-4e66-96f3-14b356502921\") " Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.279626 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tgg2\" (UniqueName: \"kubernetes.io/projected/ecfa5fcf-afae-4e66-96f3-14b356502921-kube-api-access-5tgg2\") pod \"ecfa5fcf-afae-4e66-96f3-14b356502921\" (UID: \"ecfa5fcf-afae-4e66-96f3-14b356502921\") " Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.279766 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d8702c0-4b10-41e5-b170-64a356553760-combined-ca-bundle\") pod \"9d8702c0-4b10-41e5-b170-64a356553760\" (UID: \"9d8702c0-4b10-41e5-b170-64a356553760\") " Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.279875 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d8702c0-4b10-41e5-b170-64a356553760-ovndb-tls-certs\") pod \"9d8702c0-4b10-41e5-b170-64a356553760\" (UID: \"9d8702c0-4b10-41e5-b170-64a356553760\") " Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.280503 4934 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecfa5fcf-afae-4e66-96f3-14b356502921-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.281397 4934 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ecfa5fcf-afae-4e66-96f3-14b356502921-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.284236 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecfa5fcf-afae-4e66-96f3-14b356502921-kube-api-access-5tgg2" (OuterVolumeSpecName: "kube-api-access-5tgg2") pod "ecfa5fcf-afae-4e66-96f3-14b356502921" (UID: "ecfa5fcf-afae-4e66-96f3-14b356502921"). InnerVolumeSpecName "kube-api-access-5tgg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.287004 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d8702c0-4b10-41e5-b170-64a356553760-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "9d8702c0-4b10-41e5-b170-64a356553760" (UID: "9d8702c0-4b10-41e5-b170-64a356553760"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.287048 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecfa5fcf-afae-4e66-96f3-14b356502921-scripts" (OuterVolumeSpecName: "scripts") pod "ecfa5fcf-afae-4e66-96f3-14b356502921" (UID: "ecfa5fcf-afae-4e66-96f3-14b356502921"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.287207 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d8702c0-4b10-41e5-b170-64a356553760-kube-api-access-cdbmg" (OuterVolumeSpecName: "kube-api-access-cdbmg") pod "9d8702c0-4b10-41e5-b170-64a356553760" (UID: "9d8702c0-4b10-41e5-b170-64a356553760"). InnerVolumeSpecName "kube-api-access-cdbmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.324980 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecfa5fcf-afae-4e66-96f3-14b356502921-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ecfa5fcf-afae-4e66-96f3-14b356502921" (UID: "ecfa5fcf-afae-4e66-96f3-14b356502921"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.337279 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d8702c0-4b10-41e5-b170-64a356553760-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d8702c0-4b10-41e5-b170-64a356553760" (UID: "9d8702c0-4b10-41e5-b170-64a356553760"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.349147 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d8702c0-4b10-41e5-b170-64a356553760-config" (OuterVolumeSpecName: "config") pod "9d8702c0-4b10-41e5-b170-64a356553760" (UID: "9d8702c0-4b10-41e5-b170-64a356553760"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.356464 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d8702c0-4b10-41e5-b170-64a356553760-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "9d8702c0-4b10-41e5-b170-64a356553760" (UID: "9d8702c0-4b10-41e5-b170-64a356553760"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.363036 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecfa5fcf-afae-4e66-96f3-14b356502921-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ecfa5fcf-afae-4e66-96f3-14b356502921" (UID: "ecfa5fcf-afae-4e66-96f3-14b356502921"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.384042 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d8702c0-4b10-41e5-b170-64a356553760-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.384082 4934 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d8702c0-4b10-41e5-b170-64a356553760-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.384100 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdbmg\" (UniqueName: \"kubernetes.io/projected/9d8702c0-4b10-41e5-b170-64a356553760-kube-api-access-cdbmg\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.384147 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecfa5fcf-afae-4e66-96f3-14b356502921-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.384160 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecfa5fcf-afae-4e66-96f3-14b356502921-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.384171 4934 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9d8702c0-4b10-41e5-b170-64a356553760-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.384183 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/9d8702c0-4b10-41e5-b170-64a356553760-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.384193 4934 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ecfa5fcf-afae-4e66-96f3-14b356502921-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.384204 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tgg2\" (UniqueName: \"kubernetes.io/projected/ecfa5fcf-afae-4e66-96f3-14b356502921-kube-api-access-5tgg2\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.393620 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecfa5fcf-afae-4e66-96f3-14b356502921-config-data" (OuterVolumeSpecName: "config-data") pod "ecfa5fcf-afae-4e66-96f3-14b356502921" (UID: "ecfa5fcf-afae-4e66-96f3-14b356502921"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:08:52 crc kubenswrapper[4934]: I1002 10:08:52.485955 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecfa5fcf-afae-4e66-96f3-14b356502921-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.017331 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-856f7d44fb-8lgx8" event={"ID":"9d8702c0-4b10-41e5-b170-64a356553760","Type":"ContainerDied","Data":"2a58baac49d821442c171f05a2c3c9c9971843e59c5742bee94443c034a65a7b"} Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.017353 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-856f7d44fb-8lgx8" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.017656 4934 scope.go:117] "RemoveContainer" containerID="a2462b4c5120af0b9bbf15be9bd9b839ac5c8d932d347d54f9c06988a93fd7dd" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.019068 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-vzrk9" event={"ID":"bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b","Type":"ContainerStarted","Data":"dc9d145452a9c46ec72acd9d97981007f869e0a17f76c8a91b868d88950976eb"} Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.029847 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ecfa5fcf-afae-4e66-96f3-14b356502921","Type":"ContainerDied","Data":"2398e966a7589a94b6b2a97704ab810a68ed18921a76ef565dd9c611f48190de"} Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.029919 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.047394 4934 scope.go:117] "RemoveContainer" containerID="7b1b88de67d53c8cce24ae235bd6d265be9585b8b71ccf82999a91c8aecefc42" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.076814 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-856f7d44fb-8lgx8"] Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.089927 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-856f7d44fb-8lgx8"] Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.091286 4934 scope.go:117] "RemoveContainer" containerID="6c3ed32906229b970e1b9aac959c6e9222e93273bec8a473c327763dcbfd2d93" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.103021 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-vzrk9" podStartSLOduration=2.44214111 podStartE2EDuration="13.103001074s" podCreationTimestamp="2025-10-02 10:08:40 +0000 UTC" firstStartedPulling="2025-10-02 10:08:41.233042882 +0000 UTC m=+1192.985684404" lastFinishedPulling="2025-10-02 10:08:51.893902846 +0000 UTC m=+1203.646544368" observedRunningTime="2025-10-02 10:08:53.061155279 +0000 UTC m=+1204.813796811" watchObservedRunningTime="2025-10-02 10:08:53.103001074 +0000 UTC m=+1204.855642596" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.116938 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.123460 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.124518 4934 scope.go:117] "RemoveContainer" containerID="b0a24151d342c969bcd0ce412285ed2ab87f2d65396f71b93b3f1d96e5f90ef3" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.137530 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:08:53 crc kubenswrapper[4934]: E1002 10:08:53.138245 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecfa5fcf-afae-4e66-96f3-14b356502921" containerName="sg-core" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.138372 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecfa5fcf-afae-4e66-96f3-14b356502921" containerName="sg-core" Oct 02 10:08:53 crc kubenswrapper[4934]: E1002 10:08:53.138475 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecfa5fcf-afae-4e66-96f3-14b356502921" containerName="ceilometer-notification-agent" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.138556 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecfa5fcf-afae-4e66-96f3-14b356502921" containerName="ceilometer-notification-agent" Oct 02 10:08:53 crc kubenswrapper[4934]: E1002 10:08:53.138683 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecfa5fcf-afae-4e66-96f3-14b356502921" containerName="ceilometer-central-agent" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.138761 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecfa5fcf-afae-4e66-96f3-14b356502921" containerName="ceilometer-central-agent" Oct 02 10:08:53 crc kubenswrapper[4934]: E1002 10:08:53.138848 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d8702c0-4b10-41e5-b170-64a356553760" containerName="neutron-httpd" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.138921 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d8702c0-4b10-41e5-b170-64a356553760" containerName="neutron-httpd" Oct 02 10:08:53 crc kubenswrapper[4934]: E1002 10:08:53.138997 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecfa5fcf-afae-4e66-96f3-14b356502921" containerName="proxy-httpd" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.139068 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecfa5fcf-afae-4e66-96f3-14b356502921" containerName="proxy-httpd" Oct 02 10:08:53 crc kubenswrapper[4934]: E1002 10:08:53.139156 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d8702c0-4b10-41e5-b170-64a356553760" containerName="neutron-api" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.139229 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d8702c0-4b10-41e5-b170-64a356553760" containerName="neutron-api" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.139550 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecfa5fcf-afae-4e66-96f3-14b356502921" containerName="sg-core" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.139674 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecfa5fcf-afae-4e66-96f3-14b356502921" containerName="ceilometer-notification-agent" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.139769 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d8702c0-4b10-41e5-b170-64a356553760" containerName="neutron-api" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.139865 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d8702c0-4b10-41e5-b170-64a356553760" containerName="neutron-httpd" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.139960 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecfa5fcf-afae-4e66-96f3-14b356502921" containerName="ceilometer-central-agent" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.140050 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecfa5fcf-afae-4e66-96f3-14b356502921" containerName="proxy-httpd" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.142338 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.144725 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.145128 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.145415 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.165885 4934 scope.go:117] "RemoveContainer" containerID="a5362aa5c80c6fcf6a44c0db174e8545d8e68048cb3250a039ff0ee2f1e83c03" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.198442 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06eb14a8-953f-48f9-9da1-a217d6f5568a-config-data\") pod \"ceilometer-0\" (UID: \"06eb14a8-953f-48f9-9da1-a217d6f5568a\") " pod="openstack/ceilometer-0" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.198495 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06eb14a8-953f-48f9-9da1-a217d6f5568a-run-httpd\") pod \"ceilometer-0\" (UID: \"06eb14a8-953f-48f9-9da1-a217d6f5568a\") " pod="openstack/ceilometer-0" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.198531 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/06eb14a8-953f-48f9-9da1-a217d6f5568a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"06eb14a8-953f-48f9-9da1-a217d6f5568a\") " pod="openstack/ceilometer-0" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.198552 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djv4s\" (UniqueName: \"kubernetes.io/projected/06eb14a8-953f-48f9-9da1-a217d6f5568a-kube-api-access-djv4s\") pod \"ceilometer-0\" (UID: \"06eb14a8-953f-48f9-9da1-a217d6f5568a\") " pod="openstack/ceilometer-0" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.198600 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06eb14a8-953f-48f9-9da1-a217d6f5568a-log-httpd\") pod \"ceilometer-0\" (UID: \"06eb14a8-953f-48f9-9da1-a217d6f5568a\") " pod="openstack/ceilometer-0" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.198645 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06eb14a8-953f-48f9-9da1-a217d6f5568a-scripts\") pod \"ceilometer-0\" (UID: \"06eb14a8-953f-48f9-9da1-a217d6f5568a\") " pod="openstack/ceilometer-0" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.198664 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06eb14a8-953f-48f9-9da1-a217d6f5568a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"06eb14a8-953f-48f9-9da1-a217d6f5568a\") " pod="openstack/ceilometer-0" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.200710 4934 scope.go:117] "RemoveContainer" containerID="63ed7f16e4336b15b61aea81078bf140c4e03c8913b091e80f8491b0381690a5" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.300041 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06eb14a8-953f-48f9-9da1-a217d6f5568a-config-data\") pod \"ceilometer-0\" (UID: \"06eb14a8-953f-48f9-9da1-a217d6f5568a\") " pod="openstack/ceilometer-0" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.300111 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06eb14a8-953f-48f9-9da1-a217d6f5568a-run-httpd\") pod \"ceilometer-0\" (UID: \"06eb14a8-953f-48f9-9da1-a217d6f5568a\") " pod="openstack/ceilometer-0" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.300163 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/06eb14a8-953f-48f9-9da1-a217d6f5568a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"06eb14a8-953f-48f9-9da1-a217d6f5568a\") " pod="openstack/ceilometer-0" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.300193 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djv4s\" (UniqueName: \"kubernetes.io/projected/06eb14a8-953f-48f9-9da1-a217d6f5568a-kube-api-access-djv4s\") pod \"ceilometer-0\" (UID: \"06eb14a8-953f-48f9-9da1-a217d6f5568a\") " pod="openstack/ceilometer-0" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.300231 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06eb14a8-953f-48f9-9da1-a217d6f5568a-log-httpd\") pod \"ceilometer-0\" (UID: \"06eb14a8-953f-48f9-9da1-a217d6f5568a\") " pod="openstack/ceilometer-0" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.300283 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06eb14a8-953f-48f9-9da1-a217d6f5568a-scripts\") pod \"ceilometer-0\" (UID: \"06eb14a8-953f-48f9-9da1-a217d6f5568a\") " pod="openstack/ceilometer-0" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.300306 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06eb14a8-953f-48f9-9da1-a217d6f5568a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"06eb14a8-953f-48f9-9da1-a217d6f5568a\") " pod="openstack/ceilometer-0" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.301457 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06eb14a8-953f-48f9-9da1-a217d6f5568a-run-httpd\") pod \"ceilometer-0\" (UID: \"06eb14a8-953f-48f9-9da1-a217d6f5568a\") " pod="openstack/ceilometer-0" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.301691 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06eb14a8-953f-48f9-9da1-a217d6f5568a-log-httpd\") pod \"ceilometer-0\" (UID: \"06eb14a8-953f-48f9-9da1-a217d6f5568a\") " pod="openstack/ceilometer-0" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.305105 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06eb14a8-953f-48f9-9da1-a217d6f5568a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"06eb14a8-953f-48f9-9da1-a217d6f5568a\") " pod="openstack/ceilometer-0" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.305218 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06eb14a8-953f-48f9-9da1-a217d6f5568a-config-data\") pod \"ceilometer-0\" (UID: \"06eb14a8-953f-48f9-9da1-a217d6f5568a\") " pod="openstack/ceilometer-0" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.305782 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06eb14a8-953f-48f9-9da1-a217d6f5568a-scripts\") pod \"ceilometer-0\" (UID: \"06eb14a8-953f-48f9-9da1-a217d6f5568a\") " pod="openstack/ceilometer-0" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.318363 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djv4s\" (UniqueName: \"kubernetes.io/projected/06eb14a8-953f-48f9-9da1-a217d6f5568a-kube-api-access-djv4s\") pod \"ceilometer-0\" (UID: \"06eb14a8-953f-48f9-9da1-a217d6f5568a\") " pod="openstack/ceilometer-0" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.320126 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/06eb14a8-953f-48f9-9da1-a217d6f5568a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"06eb14a8-953f-48f9-9da1-a217d6f5568a\") " pod="openstack/ceilometer-0" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.456461 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:08:53 crc kubenswrapper[4934]: I1002 10:08:53.992331 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:08:53 crc kubenswrapper[4934]: W1002 10:08:53.998072 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06eb14a8_953f_48f9_9da1_a217d6f5568a.slice/crio-fcfcec9bc4ae0b3ad4b8ffb01c18deb1a2130bd48b2e3269ad9ada9bf468b102 WatchSource:0}: Error finding container fcfcec9bc4ae0b3ad4b8ffb01c18deb1a2130bd48b2e3269ad9ada9bf468b102: Status 404 returned error can't find the container with id fcfcec9bc4ae0b3ad4b8ffb01c18deb1a2130bd48b2e3269ad9ada9bf468b102 Oct 02 10:08:54 crc kubenswrapper[4934]: I1002 10:08:54.040228 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06eb14a8-953f-48f9-9da1-a217d6f5568a","Type":"ContainerStarted","Data":"fcfcec9bc4ae0b3ad4b8ffb01c18deb1a2130bd48b2e3269ad9ada9bf468b102"} Oct 02 10:08:54 crc kubenswrapper[4934]: I1002 10:08:54.924426 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d8702c0-4b10-41e5-b170-64a356553760" path="/var/lib/kubelet/pods/9d8702c0-4b10-41e5-b170-64a356553760/volumes" Oct 02 10:08:54 crc kubenswrapper[4934]: I1002 10:08:54.926049 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecfa5fcf-afae-4e66-96f3-14b356502921" path="/var/lib/kubelet/pods/ecfa5fcf-afae-4e66-96f3-14b356502921/volumes" Oct 02 10:08:55 crc kubenswrapper[4934]: I1002 10:08:55.053239 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06eb14a8-953f-48f9-9da1-a217d6f5568a","Type":"ContainerStarted","Data":"c630cb405e8c2e671298d31e62bc1dd4507dff91d4e9081ba2703b0a2c966f1a"} Oct 02 10:08:56 crc kubenswrapper[4934]: I1002 10:08:56.065168 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06eb14a8-953f-48f9-9da1-a217d6f5568a","Type":"ContainerStarted","Data":"3a12e64befe2bba1ea2f15bb3318dd9358c70985e22da881b9157cfb0998d0a1"} Oct 02 10:08:57 crc kubenswrapper[4934]: I1002 10:08:57.078487 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06eb14a8-953f-48f9-9da1-a217d6f5568a","Type":"ContainerStarted","Data":"1e06b21ceb0c24b7ae0cae51c6becaf3172bbda2e4d646bf490f1cc858852347"} Oct 02 10:08:59 crc kubenswrapper[4934]: I1002 10:08:59.095741 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06eb14a8-953f-48f9-9da1-a217d6f5568a","Type":"ContainerStarted","Data":"1b6f098eddfd42439826b80bb65f1c06a66432a5934703606b4954bb0e943de3"} Oct 02 10:08:59 crc kubenswrapper[4934]: I1002 10:08:59.096923 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 10:08:59 crc kubenswrapper[4934]: I1002 10:08:59.118073 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.833588859 podStartE2EDuration="6.118058786s" podCreationTimestamp="2025-10-02 10:08:53 +0000 UTC" firstStartedPulling="2025-10-02 10:08:54.001955301 +0000 UTC m=+1205.754596823" lastFinishedPulling="2025-10-02 10:08:58.286425218 +0000 UTC m=+1210.039066750" observedRunningTime="2025-10-02 10:08:59.115138405 +0000 UTC m=+1210.867779927" watchObservedRunningTime="2025-10-02 10:08:59.118058786 +0000 UTC m=+1210.870700308" Oct 02 10:09:10 crc kubenswrapper[4934]: I1002 10:09:10.218871 4934 generic.go:334] "Generic (PLEG): container finished" podID="bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b" containerID="dc9d145452a9c46ec72acd9d97981007f869e0a17f76c8a91b868d88950976eb" exitCode=0 Oct 02 10:09:10 crc kubenswrapper[4934]: I1002 10:09:10.219644 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-vzrk9" event={"ID":"bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b","Type":"ContainerDied","Data":"dc9d145452a9c46ec72acd9d97981007f869e0a17f76c8a91b868d88950976eb"} Oct 02 10:09:11 crc kubenswrapper[4934]: I1002 10:09:11.712842 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-vzrk9" Oct 02 10:09:11 crc kubenswrapper[4934]: I1002 10:09:11.903906 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqzjt\" (UniqueName: \"kubernetes.io/projected/bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b-kube-api-access-hqzjt\") pod \"bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b\" (UID: \"bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b\") " Oct 02 10:09:11 crc kubenswrapper[4934]: I1002 10:09:11.904116 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b-config-data\") pod \"bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b\" (UID: \"bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b\") " Oct 02 10:09:11 crc kubenswrapper[4934]: I1002 10:09:11.904148 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b-scripts\") pod \"bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b\" (UID: \"bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b\") " Oct 02 10:09:11 crc kubenswrapper[4934]: I1002 10:09:11.904278 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b-combined-ca-bundle\") pod \"bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b\" (UID: \"bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b\") " Oct 02 10:09:11 crc kubenswrapper[4934]: I1002 10:09:11.909254 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b-scripts" (OuterVolumeSpecName: "scripts") pod "bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b" (UID: "bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:09:11 crc kubenswrapper[4934]: I1002 10:09:11.913851 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b-kube-api-access-hqzjt" (OuterVolumeSpecName: "kube-api-access-hqzjt") pod "bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b" (UID: "bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b"). InnerVolumeSpecName "kube-api-access-hqzjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:09:11 crc kubenswrapper[4934]: I1002 10:09:11.941739 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b-config-data" (OuterVolumeSpecName: "config-data") pod "bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b" (UID: "bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:09:11 crc kubenswrapper[4934]: I1002 10:09:11.945802 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b" (UID: "bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:09:12 crc kubenswrapper[4934]: I1002 10:09:12.006760 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:12 crc kubenswrapper[4934]: I1002 10:09:12.006949 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:12 crc kubenswrapper[4934]: I1002 10:09:12.007195 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:12 crc kubenswrapper[4934]: I1002 10:09:12.007317 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqzjt\" (UniqueName: \"kubernetes.io/projected/bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b-kube-api-access-hqzjt\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:12 crc kubenswrapper[4934]: I1002 10:09:12.245112 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-vzrk9" event={"ID":"bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b","Type":"ContainerDied","Data":"3234fd55176ffcc8aa8a597cff9b3be1ccb0d9f2c86c1c851f4c567c59526647"} Oct 02 10:09:12 crc kubenswrapper[4934]: I1002 10:09:12.245170 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3234fd55176ffcc8aa8a597cff9b3be1ccb0d9f2c86c1c851f4c567c59526647" Oct 02 10:09:12 crc kubenswrapper[4934]: I1002 10:09:12.245251 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-vzrk9" Oct 02 10:09:12 crc kubenswrapper[4934]: I1002 10:09:12.363697 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 10:09:12 crc kubenswrapper[4934]: E1002 10:09:12.364662 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b" containerName="nova-cell0-conductor-db-sync" Oct 02 10:09:12 crc kubenswrapper[4934]: I1002 10:09:12.364772 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b" containerName="nova-cell0-conductor-db-sync" Oct 02 10:09:12 crc kubenswrapper[4934]: I1002 10:09:12.365047 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b" containerName="nova-cell0-conductor-db-sync" Oct 02 10:09:12 crc kubenswrapper[4934]: I1002 10:09:12.365756 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 10:09:12 crc kubenswrapper[4934]: I1002 10:09:12.368122 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-96jp7" Oct 02 10:09:12 crc kubenswrapper[4934]: I1002 10:09:12.368431 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 02 10:09:12 crc kubenswrapper[4934]: I1002 10:09:12.377941 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 10:09:12 crc kubenswrapper[4934]: I1002 10:09:12.517414 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dda2730-9bb9-4113-80d3-168eb5905b2f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3dda2730-9bb9-4113-80d3-168eb5905b2f\") " pod="openstack/nova-cell0-conductor-0" Oct 02 10:09:12 crc kubenswrapper[4934]: I1002 10:09:12.517714 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dda2730-9bb9-4113-80d3-168eb5905b2f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3dda2730-9bb9-4113-80d3-168eb5905b2f\") " pod="openstack/nova-cell0-conductor-0" Oct 02 10:09:12 crc kubenswrapper[4934]: I1002 10:09:12.517752 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlnd5\" (UniqueName: \"kubernetes.io/projected/3dda2730-9bb9-4113-80d3-168eb5905b2f-kube-api-access-rlnd5\") pod \"nova-cell0-conductor-0\" (UID: \"3dda2730-9bb9-4113-80d3-168eb5905b2f\") " pod="openstack/nova-cell0-conductor-0" Oct 02 10:09:12 crc kubenswrapper[4934]: I1002 10:09:12.620027 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dda2730-9bb9-4113-80d3-168eb5905b2f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3dda2730-9bb9-4113-80d3-168eb5905b2f\") " pod="openstack/nova-cell0-conductor-0" Oct 02 10:09:12 crc kubenswrapper[4934]: I1002 10:09:12.620067 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlnd5\" (UniqueName: \"kubernetes.io/projected/3dda2730-9bb9-4113-80d3-168eb5905b2f-kube-api-access-rlnd5\") pod \"nova-cell0-conductor-0\" (UID: \"3dda2730-9bb9-4113-80d3-168eb5905b2f\") " pod="openstack/nova-cell0-conductor-0" Oct 02 10:09:12 crc kubenswrapper[4934]: I1002 10:09:12.620156 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dda2730-9bb9-4113-80d3-168eb5905b2f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3dda2730-9bb9-4113-80d3-168eb5905b2f\") " pod="openstack/nova-cell0-conductor-0" Oct 02 10:09:12 crc kubenswrapper[4934]: I1002 10:09:12.626766 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dda2730-9bb9-4113-80d3-168eb5905b2f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3dda2730-9bb9-4113-80d3-168eb5905b2f\") " pod="openstack/nova-cell0-conductor-0" Oct 02 10:09:12 crc kubenswrapper[4934]: I1002 10:09:12.637871 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dda2730-9bb9-4113-80d3-168eb5905b2f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3dda2730-9bb9-4113-80d3-168eb5905b2f\") " pod="openstack/nova-cell0-conductor-0" Oct 02 10:09:12 crc kubenswrapper[4934]: I1002 10:09:12.638951 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlnd5\" (UniqueName: \"kubernetes.io/projected/3dda2730-9bb9-4113-80d3-168eb5905b2f-kube-api-access-rlnd5\") pod \"nova-cell0-conductor-0\" (UID: \"3dda2730-9bb9-4113-80d3-168eb5905b2f\") " pod="openstack/nova-cell0-conductor-0" Oct 02 10:09:12 crc kubenswrapper[4934]: I1002 10:09:12.681374 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 10:09:13 crc kubenswrapper[4934]: I1002 10:09:13.176599 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 10:09:13 crc kubenswrapper[4934]: I1002 10:09:13.260002 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"3dda2730-9bb9-4113-80d3-168eb5905b2f","Type":"ContainerStarted","Data":"397bdf51f4e9d02cbc9b0c8f8eee6ce13158d4ffeeb8f8a82b3db7867e57f59b"} Oct 02 10:09:14 crc kubenswrapper[4934]: I1002 10:09:14.274241 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"3dda2730-9bb9-4113-80d3-168eb5905b2f","Type":"ContainerStarted","Data":"75db33a0a204cb84036a785aefe5e3a861f5bb1c02c563e44111dc7a54a02cb7"} Oct 02 10:09:14 crc kubenswrapper[4934]: I1002 10:09:14.274660 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 02 10:09:22 crc kubenswrapper[4934]: I1002 10:09:22.724146 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 02 10:09:22 crc kubenswrapper[4934]: I1002 10:09:22.745257 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=10.745232306 podStartE2EDuration="10.745232306s" podCreationTimestamp="2025-10-02 10:09:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:09:14.300228495 +0000 UTC m=+1226.052870067" watchObservedRunningTime="2025-10-02 10:09:22.745232306 +0000 UTC m=+1234.497873868" Oct 02 10:09:23 crc kubenswrapper[4934]: I1002 10:09:23.539160 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.318886 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-zkqqc"] Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.321233 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zkqqc" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.324309 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.324566 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.337068 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-zkqqc"] Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.463520 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed820a81-72fc-4d84-876f-9def630d4ff7-config-data\") pod \"nova-cell0-cell-mapping-zkqqc\" (UID: \"ed820a81-72fc-4d84-876f-9def630d4ff7\") " pod="openstack/nova-cell0-cell-mapping-zkqqc" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.463728 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xj9ff\" (UniqueName: \"kubernetes.io/projected/ed820a81-72fc-4d84-876f-9def630d4ff7-kube-api-access-xj9ff\") pod \"nova-cell0-cell-mapping-zkqqc\" (UID: \"ed820a81-72fc-4d84-876f-9def630d4ff7\") " pod="openstack/nova-cell0-cell-mapping-zkqqc" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.463784 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed820a81-72fc-4d84-876f-9def630d4ff7-scripts\") pod \"nova-cell0-cell-mapping-zkqqc\" (UID: \"ed820a81-72fc-4d84-876f-9def630d4ff7\") " pod="openstack/nova-cell0-cell-mapping-zkqqc" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.463850 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed820a81-72fc-4d84-876f-9def630d4ff7-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zkqqc\" (UID: \"ed820a81-72fc-4d84-876f-9def630d4ff7\") " pod="openstack/nova-cell0-cell-mapping-zkqqc" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.464140 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6c8h9"] Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.465299 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-6c8h9" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.467204 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.479475 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.527467 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6c8h9"] Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.565200 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkd92\" (UniqueName: \"kubernetes.io/projected/490173ef-a363-4b84-8802-6a69bb9bdb76-kube-api-access-qkd92\") pod \"nova-cell1-conductor-db-sync-6c8h9\" (UID: \"490173ef-a363-4b84-8802-6a69bb9bdb76\") " pod="openstack/nova-cell1-conductor-db-sync-6c8h9" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.565275 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xj9ff\" (UniqueName: \"kubernetes.io/projected/ed820a81-72fc-4d84-876f-9def630d4ff7-kube-api-access-xj9ff\") pod \"nova-cell0-cell-mapping-zkqqc\" (UID: \"ed820a81-72fc-4d84-876f-9def630d4ff7\") " pod="openstack/nova-cell0-cell-mapping-zkqqc" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.565320 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/490173ef-a363-4b84-8802-6a69bb9bdb76-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-6c8h9\" (UID: \"490173ef-a363-4b84-8802-6a69bb9bdb76\") " pod="openstack/nova-cell1-conductor-db-sync-6c8h9" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.565353 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed820a81-72fc-4d84-876f-9def630d4ff7-scripts\") pod \"nova-cell0-cell-mapping-zkqqc\" (UID: \"ed820a81-72fc-4d84-876f-9def630d4ff7\") " pod="openstack/nova-cell0-cell-mapping-zkqqc" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.565405 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed820a81-72fc-4d84-876f-9def630d4ff7-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zkqqc\" (UID: \"ed820a81-72fc-4d84-876f-9def630d4ff7\") " pod="openstack/nova-cell0-cell-mapping-zkqqc" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.565444 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/490173ef-a363-4b84-8802-6a69bb9bdb76-scripts\") pod \"nova-cell1-conductor-db-sync-6c8h9\" (UID: \"490173ef-a363-4b84-8802-6a69bb9bdb76\") " pod="openstack/nova-cell1-conductor-db-sync-6c8h9" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.565502 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed820a81-72fc-4d84-876f-9def630d4ff7-config-data\") pod \"nova-cell0-cell-mapping-zkqqc\" (UID: \"ed820a81-72fc-4d84-876f-9def630d4ff7\") " pod="openstack/nova-cell0-cell-mapping-zkqqc" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.565763 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/490173ef-a363-4b84-8802-6a69bb9bdb76-config-data\") pod \"nova-cell1-conductor-db-sync-6c8h9\" (UID: \"490173ef-a363-4b84-8802-6a69bb9bdb76\") " pod="openstack/nova-cell1-conductor-db-sync-6c8h9" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.572455 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed820a81-72fc-4d84-876f-9def630d4ff7-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zkqqc\" (UID: \"ed820a81-72fc-4d84-876f-9def630d4ff7\") " pod="openstack/nova-cell0-cell-mapping-zkqqc" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.572482 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed820a81-72fc-4d84-876f-9def630d4ff7-config-data\") pod \"nova-cell0-cell-mapping-zkqqc\" (UID: \"ed820a81-72fc-4d84-876f-9def630d4ff7\") " pod="openstack/nova-cell0-cell-mapping-zkqqc" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.579120 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed820a81-72fc-4d84-876f-9def630d4ff7-scripts\") pod \"nova-cell0-cell-mapping-zkqqc\" (UID: \"ed820a81-72fc-4d84-876f-9def630d4ff7\") " pod="openstack/nova-cell0-cell-mapping-zkqqc" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.588076 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.589470 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.592093 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xj9ff\" (UniqueName: \"kubernetes.io/projected/ed820a81-72fc-4d84-876f-9def630d4ff7-kube-api-access-xj9ff\") pod \"nova-cell0-cell-mapping-zkqqc\" (UID: \"ed820a81-72fc-4d84-876f-9def630d4ff7\") " pod="openstack/nova-cell0-cell-mapping-zkqqc" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.592628 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.611624 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.653238 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zkqqc" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.674945 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/490173ef-a363-4b84-8802-6a69bb9bdb76-scripts\") pod \"nova-cell1-conductor-db-sync-6c8h9\" (UID: \"490173ef-a363-4b84-8802-6a69bb9bdb76\") " pod="openstack/nova-cell1-conductor-db-sync-6c8h9" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.675035 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8fqf\" (UniqueName: \"kubernetes.io/projected/21e92017-966a-499d-8bb5-3225aacd6dd7-kube-api-access-l8fqf\") pod \"nova-cell1-novncproxy-0\" (UID: \"21e92017-966a-499d-8bb5-3225aacd6dd7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.675075 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/490173ef-a363-4b84-8802-6a69bb9bdb76-config-data\") pod \"nova-cell1-conductor-db-sync-6c8h9\" (UID: \"490173ef-a363-4b84-8802-6a69bb9bdb76\") " pod="openstack/nova-cell1-conductor-db-sync-6c8h9" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.675131 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkd92\" (UniqueName: \"kubernetes.io/projected/490173ef-a363-4b84-8802-6a69bb9bdb76-kube-api-access-qkd92\") pod \"nova-cell1-conductor-db-sync-6c8h9\" (UID: \"490173ef-a363-4b84-8802-6a69bb9bdb76\") " pod="openstack/nova-cell1-conductor-db-sync-6c8h9" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.675168 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e92017-966a-499d-8bb5-3225aacd6dd7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"21e92017-966a-499d-8bb5-3225aacd6dd7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.675210 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/490173ef-a363-4b84-8802-6a69bb9bdb76-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-6c8h9\" (UID: \"490173ef-a363-4b84-8802-6a69bb9bdb76\") " pod="openstack/nova-cell1-conductor-db-sync-6c8h9" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.675249 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e92017-966a-499d-8bb5-3225aacd6dd7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"21e92017-966a-499d-8bb5-3225aacd6dd7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.678677 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.679891 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.682959 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.684477 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/490173ef-a363-4b84-8802-6a69bb9bdb76-config-data\") pod \"nova-cell1-conductor-db-sync-6c8h9\" (UID: \"490173ef-a363-4b84-8802-6a69bb9bdb76\") " pod="openstack/nova-cell1-conductor-db-sync-6c8h9" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.686310 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/490173ef-a363-4b84-8802-6a69bb9bdb76-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-6c8h9\" (UID: \"490173ef-a363-4b84-8802-6a69bb9bdb76\") " pod="openstack/nova-cell1-conductor-db-sync-6c8h9" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.689124 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/490173ef-a363-4b84-8802-6a69bb9bdb76-scripts\") pod \"nova-cell1-conductor-db-sync-6c8h9\" (UID: \"490173ef-a363-4b84-8802-6a69bb9bdb76\") " pod="openstack/nova-cell1-conductor-db-sync-6c8h9" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.740094 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkd92\" (UniqueName: \"kubernetes.io/projected/490173ef-a363-4b84-8802-6a69bb9bdb76-kube-api-access-qkd92\") pod \"nova-cell1-conductor-db-sync-6c8h9\" (UID: \"490173ef-a363-4b84-8802-6a69bb9bdb76\") " pod="openstack/nova-cell1-conductor-db-sync-6c8h9" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.740166 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.741921 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.746053 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.777771 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.780204 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e92017-966a-499d-8bb5-3225aacd6dd7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"21e92017-966a-499d-8bb5-3225aacd6dd7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.780337 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8fqf\" (UniqueName: \"kubernetes.io/projected/21e92017-966a-499d-8bb5-3225aacd6dd7-kube-api-access-l8fqf\") pod \"nova-cell1-novncproxy-0\" (UID: \"21e92017-966a-499d-8bb5-3225aacd6dd7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.780374 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/004218d8-62bb-470f-b184-67d0ac97a27c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"004218d8-62bb-470f-b184-67d0ac97a27c\") " pod="openstack/nova-scheduler-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.780419 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e92017-966a-499d-8bb5-3225aacd6dd7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"21e92017-966a-499d-8bb5-3225aacd6dd7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.780437 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/004218d8-62bb-470f-b184-67d0ac97a27c-config-data\") pod \"nova-scheduler-0\" (UID: \"004218d8-62bb-470f-b184-67d0ac97a27c\") " pod="openstack/nova-scheduler-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.780462 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkdm2\" (UniqueName: \"kubernetes.io/projected/004218d8-62bb-470f-b184-67d0ac97a27c-kube-api-access-pkdm2\") pod \"nova-scheduler-0\" (UID: \"004218d8-62bb-470f-b184-67d0ac97a27c\") " pod="openstack/nova-scheduler-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.782156 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.784011 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.786121 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e92017-966a-499d-8bb5-3225aacd6dd7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"21e92017-966a-499d-8bb5-3225aacd6dd7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.786508 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-6c8h9" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.794067 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e92017-966a-499d-8bb5-3225aacd6dd7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"21e92017-966a-499d-8bb5-3225aacd6dd7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.799125 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.809210 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8fqf\" (UniqueName: \"kubernetes.io/projected/21e92017-966a-499d-8bb5-3225aacd6dd7-kube-api-access-l8fqf\") pod \"nova-cell1-novncproxy-0\" (UID: \"21e92017-966a-499d-8bb5-3225aacd6dd7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.809669 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.821995 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.849816 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.881914 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cb6eb34-3995-49cb-b6a4-afeb25641e6a-config-data\") pod \"nova-metadata-0\" (UID: \"7cb6eb34-3995-49cb-b6a4-afeb25641e6a\") " pod="openstack/nova-metadata-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.881959 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4da04f7-1f59-4acc-90e0-7b7cba140393-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b4da04f7-1f59-4acc-90e0-7b7cba140393\") " pod="openstack/nova-api-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.881997 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/004218d8-62bb-470f-b184-67d0ac97a27c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"004218d8-62bb-470f-b184-67d0ac97a27c\") " pod="openstack/nova-scheduler-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.882032 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlb44\" (UniqueName: \"kubernetes.io/projected/b4da04f7-1f59-4acc-90e0-7b7cba140393-kube-api-access-vlb44\") pod \"nova-api-0\" (UID: \"b4da04f7-1f59-4acc-90e0-7b7cba140393\") " pod="openstack/nova-api-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.882060 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/004218d8-62bb-470f-b184-67d0ac97a27c-config-data\") pod \"nova-scheduler-0\" (UID: \"004218d8-62bb-470f-b184-67d0ac97a27c\") " pod="openstack/nova-scheduler-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.882146 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cb6eb34-3995-49cb-b6a4-afeb25641e6a-logs\") pod \"nova-metadata-0\" (UID: \"7cb6eb34-3995-49cb-b6a4-afeb25641e6a\") " pod="openstack/nova-metadata-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.882168 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wx498\" (UniqueName: \"kubernetes.io/projected/7cb6eb34-3995-49cb-b6a4-afeb25641e6a-kube-api-access-wx498\") pod \"nova-metadata-0\" (UID: \"7cb6eb34-3995-49cb-b6a4-afeb25641e6a\") " pod="openstack/nova-metadata-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.882212 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4da04f7-1f59-4acc-90e0-7b7cba140393-logs\") pod \"nova-api-0\" (UID: \"b4da04f7-1f59-4acc-90e0-7b7cba140393\") " pod="openstack/nova-api-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.882237 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkdm2\" (UniqueName: \"kubernetes.io/projected/004218d8-62bb-470f-b184-67d0ac97a27c-kube-api-access-pkdm2\") pod \"nova-scheduler-0\" (UID: \"004218d8-62bb-470f-b184-67d0ac97a27c\") " pod="openstack/nova-scheduler-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.882278 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cb6eb34-3995-49cb-b6a4-afeb25641e6a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7cb6eb34-3995-49cb-b6a4-afeb25641e6a\") " pod="openstack/nova-metadata-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.882303 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4da04f7-1f59-4acc-90e0-7b7cba140393-config-data\") pod \"nova-api-0\" (UID: \"b4da04f7-1f59-4acc-90e0-7b7cba140393\") " pod="openstack/nova-api-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.882853 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b5cc5d6f9-v9lw6"] Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.889024 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.899790 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/004218d8-62bb-470f-b184-67d0ac97a27c-config-data\") pod \"nova-scheduler-0\" (UID: \"004218d8-62bb-470f-b184-67d0ac97a27c\") " pod="openstack/nova-scheduler-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.901132 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/004218d8-62bb-470f-b184-67d0ac97a27c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"004218d8-62bb-470f-b184-67d0ac97a27c\") " pod="openstack/nova-scheduler-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.905518 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b5cc5d6f9-v9lw6"] Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.913199 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkdm2\" (UniqueName: \"kubernetes.io/projected/004218d8-62bb-470f-b184-67d0ac97a27c-kube-api-access-pkdm2\") pod \"nova-scheduler-0\" (UID: \"004218d8-62bb-470f-b184-67d0ac97a27c\") " pod="openstack/nova-scheduler-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.992466 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-ovsdbserver-nb\") pod \"dnsmasq-dns-b5cc5d6f9-v9lw6\" (UID: \"a446ab5e-52d1-4962-a383-3fab4a5791f7\") " pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.992592 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlb44\" (UniqueName: \"kubernetes.io/projected/b4da04f7-1f59-4acc-90e0-7b7cba140393-kube-api-access-vlb44\") pod \"nova-api-0\" (UID: \"b4da04f7-1f59-4acc-90e0-7b7cba140393\") " pod="openstack/nova-api-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.992657 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-ovsdbserver-sb\") pod \"dnsmasq-dns-b5cc5d6f9-v9lw6\" (UID: \"a446ab5e-52d1-4962-a383-3fab4a5791f7\") " pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.992719 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cb6eb34-3995-49cb-b6a4-afeb25641e6a-logs\") pod \"nova-metadata-0\" (UID: \"7cb6eb34-3995-49cb-b6a4-afeb25641e6a\") " pod="openstack/nova-metadata-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.992746 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wx498\" (UniqueName: \"kubernetes.io/projected/7cb6eb34-3995-49cb-b6a4-afeb25641e6a-kube-api-access-wx498\") pod \"nova-metadata-0\" (UID: \"7cb6eb34-3995-49cb-b6a4-afeb25641e6a\") " pod="openstack/nova-metadata-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.992775 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4da04f7-1f59-4acc-90e0-7b7cba140393-logs\") pod \"nova-api-0\" (UID: \"b4da04f7-1f59-4acc-90e0-7b7cba140393\") " pod="openstack/nova-api-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.992828 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cb6eb34-3995-49cb-b6a4-afeb25641e6a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7cb6eb34-3995-49cb-b6a4-afeb25641e6a\") " pod="openstack/nova-metadata-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.992878 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4da04f7-1f59-4acc-90e0-7b7cba140393-config-data\") pod \"nova-api-0\" (UID: \"b4da04f7-1f59-4acc-90e0-7b7cba140393\") " pod="openstack/nova-api-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.992957 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-config\") pod \"dnsmasq-dns-b5cc5d6f9-v9lw6\" (UID: \"a446ab5e-52d1-4962-a383-3fab4a5791f7\") " pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.993122 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-dns-swift-storage-0\") pod \"dnsmasq-dns-b5cc5d6f9-v9lw6\" (UID: \"a446ab5e-52d1-4962-a383-3fab4a5791f7\") " pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.993145 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cb6eb34-3995-49cb-b6a4-afeb25641e6a-config-data\") pod \"nova-metadata-0\" (UID: \"7cb6eb34-3995-49cb-b6a4-afeb25641e6a\") " pod="openstack/nova-metadata-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.993167 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-dns-svc\") pod \"dnsmasq-dns-b5cc5d6f9-v9lw6\" (UID: \"a446ab5e-52d1-4962-a383-3fab4a5791f7\") " pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.993219 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4da04f7-1f59-4acc-90e0-7b7cba140393-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b4da04f7-1f59-4acc-90e0-7b7cba140393\") " pod="openstack/nova-api-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.993269 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhxjk\" (UniqueName: \"kubernetes.io/projected/a446ab5e-52d1-4962-a383-3fab4a5791f7-kube-api-access-dhxjk\") pod \"dnsmasq-dns-b5cc5d6f9-v9lw6\" (UID: \"a446ab5e-52d1-4962-a383-3fab4a5791f7\") " pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.993350 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cb6eb34-3995-49cb-b6a4-afeb25641e6a-logs\") pod \"nova-metadata-0\" (UID: \"7cb6eb34-3995-49cb-b6a4-afeb25641e6a\") " pod="openstack/nova-metadata-0" Oct 02 10:09:24 crc kubenswrapper[4934]: I1002 10:09:24.994823 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4da04f7-1f59-4acc-90e0-7b7cba140393-logs\") pod \"nova-api-0\" (UID: \"b4da04f7-1f59-4acc-90e0-7b7cba140393\") " pod="openstack/nova-api-0" Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.004103 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cb6eb34-3995-49cb-b6a4-afeb25641e6a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7cb6eb34-3995-49cb-b6a4-afeb25641e6a\") " pod="openstack/nova-metadata-0" Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.010433 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cb6eb34-3995-49cb-b6a4-afeb25641e6a-config-data\") pod \"nova-metadata-0\" (UID: \"7cb6eb34-3995-49cb-b6a4-afeb25641e6a\") " pod="openstack/nova-metadata-0" Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.010520 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4da04f7-1f59-4acc-90e0-7b7cba140393-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b4da04f7-1f59-4acc-90e0-7b7cba140393\") " pod="openstack/nova-api-0" Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.012393 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlb44\" (UniqueName: \"kubernetes.io/projected/b4da04f7-1f59-4acc-90e0-7b7cba140393-kube-api-access-vlb44\") pod \"nova-api-0\" (UID: \"b4da04f7-1f59-4acc-90e0-7b7cba140393\") " pod="openstack/nova-api-0" Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.012831 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wx498\" (UniqueName: \"kubernetes.io/projected/7cb6eb34-3995-49cb-b6a4-afeb25641e6a-kube-api-access-wx498\") pod \"nova-metadata-0\" (UID: \"7cb6eb34-3995-49cb-b6a4-afeb25641e6a\") " pod="openstack/nova-metadata-0" Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.013560 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4da04f7-1f59-4acc-90e0-7b7cba140393-config-data\") pod \"nova-api-0\" (UID: \"b4da04f7-1f59-4acc-90e0-7b7cba140393\") " pod="openstack/nova-api-0" Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.097688 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-config\") pod \"dnsmasq-dns-b5cc5d6f9-v9lw6\" (UID: \"a446ab5e-52d1-4962-a383-3fab4a5791f7\") " pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.097796 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-dns-swift-storage-0\") pod \"dnsmasq-dns-b5cc5d6f9-v9lw6\" (UID: \"a446ab5e-52d1-4962-a383-3fab4a5791f7\") " pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.097817 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-dns-svc\") pod \"dnsmasq-dns-b5cc5d6f9-v9lw6\" (UID: \"a446ab5e-52d1-4962-a383-3fab4a5791f7\") " pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.097855 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhxjk\" (UniqueName: \"kubernetes.io/projected/a446ab5e-52d1-4962-a383-3fab4a5791f7-kube-api-access-dhxjk\") pod \"dnsmasq-dns-b5cc5d6f9-v9lw6\" (UID: \"a446ab5e-52d1-4962-a383-3fab4a5791f7\") " pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.097893 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-ovsdbserver-nb\") pod \"dnsmasq-dns-b5cc5d6f9-v9lw6\" (UID: \"a446ab5e-52d1-4962-a383-3fab4a5791f7\") " pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.097927 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-ovsdbserver-sb\") pod \"dnsmasq-dns-b5cc5d6f9-v9lw6\" (UID: \"a446ab5e-52d1-4962-a383-3fab4a5791f7\") " pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.098816 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-ovsdbserver-sb\") pod \"dnsmasq-dns-b5cc5d6f9-v9lw6\" (UID: \"a446ab5e-52d1-4962-a383-3fab4a5791f7\") " pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.099401 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-dns-swift-storage-0\") pod \"dnsmasq-dns-b5cc5d6f9-v9lw6\" (UID: \"a446ab5e-52d1-4962-a383-3fab4a5791f7\") " pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.099494 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-dns-svc\") pod \"dnsmasq-dns-b5cc5d6f9-v9lw6\" (UID: \"a446ab5e-52d1-4962-a383-3fab4a5791f7\") " pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.100053 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-config\") pod \"dnsmasq-dns-b5cc5d6f9-v9lw6\" (UID: \"a446ab5e-52d1-4962-a383-3fab4a5791f7\") " pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.100397 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-ovsdbserver-nb\") pod \"dnsmasq-dns-b5cc5d6f9-v9lw6\" (UID: \"a446ab5e-52d1-4962-a383-3fab4a5791f7\") " pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.119058 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhxjk\" (UniqueName: \"kubernetes.io/projected/a446ab5e-52d1-4962-a383-3fab4a5791f7-kube-api-access-dhxjk\") pod \"dnsmasq-dns-b5cc5d6f9-v9lw6\" (UID: \"a446ab5e-52d1-4962-a383-3fab4a5791f7\") " pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.174676 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.210071 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.248761 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.268520 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.352567 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-zkqqc"] Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.397471 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6c8h9"] Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.419520 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zkqqc" event={"ID":"ed820a81-72fc-4d84-876f-9def630d4ff7","Type":"ContainerStarted","Data":"a1c339ab77a926bc8a5fc8a29f16b23b5fc3fe605689c09863b2cd3f4e9142c5"} Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.496631 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.704154 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.822263 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:09:25 crc kubenswrapper[4934]: W1002 10:09:25.843972 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7cb6eb34_3995_49cb_b6a4_afeb25641e6a.slice/crio-dd7866576be4b5cf7f5de0102b5c7667c430898b16a4aa2b636e0b581019d815 WatchSource:0}: Error finding container dd7866576be4b5cf7f5de0102b5c7667c430898b16a4aa2b636e0b581019d815: Status 404 returned error can't find the container with id dd7866576be4b5cf7f5de0102b5c7667c430898b16a4aa2b636e0b581019d815 Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.848481 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:09:25 crc kubenswrapper[4934]: I1002 10:09:25.989634 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b5cc5d6f9-v9lw6"] Oct 02 10:09:26 crc kubenswrapper[4934]: I1002 10:09:26.427877 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7cb6eb34-3995-49cb-b6a4-afeb25641e6a","Type":"ContainerStarted","Data":"dd7866576be4b5cf7f5de0102b5c7667c430898b16a4aa2b636e0b581019d815"} Oct 02 10:09:26 crc kubenswrapper[4934]: I1002 10:09:26.429817 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"21e92017-966a-499d-8bb5-3225aacd6dd7","Type":"ContainerStarted","Data":"7766f03265fb0fd733e4d2139d3d9e86739a072eb857d065a3c448ee5bb64cdc"} Oct 02 10:09:26 crc kubenswrapper[4934]: I1002 10:09:26.431212 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zkqqc" event={"ID":"ed820a81-72fc-4d84-876f-9def630d4ff7","Type":"ContainerStarted","Data":"8d3c9f48dcfcd16e834077a1b8cefb91be17eaec6d3b75578d4108c226272f38"} Oct 02 10:09:26 crc kubenswrapper[4934]: I1002 10:09:26.432472 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"004218d8-62bb-470f-b184-67d0ac97a27c","Type":"ContainerStarted","Data":"1632b8483ea033cbbba2c7c067242ff1fd2f18a267115927a0e9d79df8b1f449"} Oct 02 10:09:26 crc kubenswrapper[4934]: I1002 10:09:26.433665 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b4da04f7-1f59-4acc-90e0-7b7cba140393","Type":"ContainerStarted","Data":"ddc17ca02bee1b3765b3561124fc12110839ea616bc262bfe97657e44dbf103c"} Oct 02 10:09:26 crc kubenswrapper[4934]: I1002 10:09:26.435353 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-6c8h9" event={"ID":"490173ef-a363-4b84-8802-6a69bb9bdb76","Type":"ContainerStarted","Data":"6b77fc0e555402c151a111643d602de419a5a7a8fe3ab9b1c50acb80d1af507a"} Oct 02 10:09:26 crc kubenswrapper[4934]: I1002 10:09:26.435384 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-6c8h9" event={"ID":"490173ef-a363-4b84-8802-6a69bb9bdb76","Type":"ContainerStarted","Data":"6bd88e6b84c85c775d5e2aa96af34b9bcb2b5a2ea400063eb4cf3835e68f5b28"} Oct 02 10:09:26 crc kubenswrapper[4934]: I1002 10:09:26.436412 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" event={"ID":"a446ab5e-52d1-4962-a383-3fab4a5791f7","Type":"ContainerStarted","Data":"fa8a18ff773cb7d961fb61e8ce3bd4c70f596c8aefd2d332d278a31ca1093c06"} Oct 02 10:09:27 crc kubenswrapper[4934]: I1002 10:09:27.446334 4934 generic.go:334] "Generic (PLEG): container finished" podID="a446ab5e-52d1-4962-a383-3fab4a5791f7" containerID="9bbf64eacc67029ce319bc27321ff52fbff4672a4c87120bb01bfa35df7f1b98" exitCode=0 Oct 02 10:09:27 crc kubenswrapper[4934]: I1002 10:09:27.446537 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" event={"ID":"a446ab5e-52d1-4962-a383-3fab4a5791f7","Type":"ContainerDied","Data":"9bbf64eacc67029ce319bc27321ff52fbff4672a4c87120bb01bfa35df7f1b98"} Oct 02 10:09:27 crc kubenswrapper[4934]: I1002 10:09:27.472520 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-zkqqc" podStartSLOduration=3.472497318 podStartE2EDuration="3.472497318s" podCreationTimestamp="2025-10-02 10:09:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:09:27.462402329 +0000 UTC m=+1239.215043871" watchObservedRunningTime="2025-10-02 10:09:27.472497318 +0000 UTC m=+1239.225138850" Oct 02 10:09:27 crc kubenswrapper[4934]: I1002 10:09:27.507092 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-6c8h9" podStartSLOduration=3.507074672 podStartE2EDuration="3.507074672s" podCreationTimestamp="2025-10-02 10:09:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:09:27.49794311 +0000 UTC m=+1239.250584622" watchObservedRunningTime="2025-10-02 10:09:27.507074672 +0000 UTC m=+1239.259716194" Oct 02 10:09:28 crc kubenswrapper[4934]: I1002 10:09:28.475037 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" event={"ID":"a446ab5e-52d1-4962-a383-3fab4a5791f7","Type":"ContainerStarted","Data":"87ed1325186219049d1c236d9ed445e9035d32d61ebb8891b0fff2e50ae2eb96"} Oct 02 10:09:28 crc kubenswrapper[4934]: I1002 10:09:28.477212 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" Oct 02 10:09:28 crc kubenswrapper[4934]: I1002 10:09:28.556287 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" podStartSLOduration=4.556266376 podStartE2EDuration="4.556266376s" podCreationTimestamp="2025-10-02 10:09:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:09:28.522775002 +0000 UTC m=+1240.275416534" watchObservedRunningTime="2025-10-02 10:09:28.556266376 +0000 UTC m=+1240.308907898" Oct 02 10:09:28 crc kubenswrapper[4934]: I1002 10:09:28.558568 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:09:28 crc kubenswrapper[4934]: I1002 10:09:28.567810 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 10:09:31 crc kubenswrapper[4934]: I1002 10:09:31.450890 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 10:09:31 crc kubenswrapper[4934]: I1002 10:09:31.451455 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="0d5014eb-503f-45ea-bb78-bc5978dbaa69" containerName="kube-state-metrics" containerID="cri-o://caccc0b41d120be902b372c1c34ce974b7a351bec8cbcf762fb7d2fb779776dd" gracePeriod=30 Oct 02 10:09:31 crc kubenswrapper[4934]: I1002 10:09:31.516217 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7cb6eb34-3995-49cb-b6a4-afeb25641e6a","Type":"ContainerStarted","Data":"7f84fb7c6e9e5d88dc364e7a4dfa9e56d0e2b98f171280f1351177ac7b560759"} Oct 02 10:09:31 crc kubenswrapper[4934]: I1002 10:09:31.516263 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7cb6eb34-3995-49cb-b6a4-afeb25641e6a","Type":"ContainerStarted","Data":"b077cf2ec73538e90e92ac14e7c3ac51e222f05cddaeb92375d89df43e635b1e"} Oct 02 10:09:31 crc kubenswrapper[4934]: I1002 10:09:31.516381 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7cb6eb34-3995-49cb-b6a4-afeb25641e6a" containerName="nova-metadata-log" containerID="cri-o://b077cf2ec73538e90e92ac14e7c3ac51e222f05cddaeb92375d89df43e635b1e" gracePeriod=30 Oct 02 10:09:31 crc kubenswrapper[4934]: I1002 10:09:31.516465 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7cb6eb34-3995-49cb-b6a4-afeb25641e6a" containerName="nova-metadata-metadata" containerID="cri-o://7f84fb7c6e9e5d88dc364e7a4dfa9e56d0e2b98f171280f1351177ac7b560759" gracePeriod=30 Oct 02 10:09:31 crc kubenswrapper[4934]: I1002 10:09:31.531863 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"21e92017-966a-499d-8bb5-3225aacd6dd7","Type":"ContainerStarted","Data":"494857a62fbea1b42d73d034dc76d537a8a6f515d0153bc68199f0984a52453d"} Oct 02 10:09:31 crc kubenswrapper[4934]: I1002 10:09:31.532000 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="21e92017-966a-499d-8bb5-3225aacd6dd7" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://494857a62fbea1b42d73d034dc76d537a8a6f515d0153bc68199f0984a52453d" gracePeriod=30 Oct 02 10:09:31 crc kubenswrapper[4934]: I1002 10:09:31.542451 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"004218d8-62bb-470f-b184-67d0ac97a27c","Type":"ContainerStarted","Data":"64480ebe56c42e093c07f270ecff0f024eb20a7d01a5f42839131f3edf140a1d"} Oct 02 10:09:31 crc kubenswrapper[4934]: I1002 10:09:31.543324 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.161962594 podStartE2EDuration="7.543307076s" podCreationTimestamp="2025-10-02 10:09:24 +0000 UTC" firstStartedPulling="2025-10-02 10:09:25.848760212 +0000 UTC m=+1237.601401734" lastFinishedPulling="2025-10-02 10:09:30.230104704 +0000 UTC m=+1241.982746216" observedRunningTime="2025-10-02 10:09:31.540135739 +0000 UTC m=+1243.292777271" watchObservedRunningTime="2025-10-02 10:09:31.543307076 +0000 UTC m=+1243.295948588" Oct 02 10:09:31 crc kubenswrapper[4934]: I1002 10:09:31.552614 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b4da04f7-1f59-4acc-90e0-7b7cba140393","Type":"ContainerStarted","Data":"a6d1113b52ffb7001a8b2d59c298de5526ba2211698784ccbcdf6bb49e0e2e01"} Oct 02 10:09:31 crc kubenswrapper[4934]: I1002 10:09:31.552657 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b4da04f7-1f59-4acc-90e0-7b7cba140393","Type":"ContainerStarted","Data":"2b6fe7b50c803ac8ece2433dd06ec23dcfef9b8a9b399c435d0bd26a0fc0e9f2"} Oct 02 10:09:31 crc kubenswrapper[4934]: I1002 10:09:31.580459 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.882699036 podStartE2EDuration="7.580442821s" podCreationTimestamp="2025-10-02 10:09:24 +0000 UTC" firstStartedPulling="2025-10-02 10:09:25.530332063 +0000 UTC m=+1237.282973605" lastFinishedPulling="2025-10-02 10:09:30.228075868 +0000 UTC m=+1241.980717390" observedRunningTime="2025-10-02 10:09:31.574999111 +0000 UTC m=+1243.327640633" watchObservedRunningTime="2025-10-02 10:09:31.580442821 +0000 UTC m=+1243.333084343" Oct 02 10:09:31 crc kubenswrapper[4934]: I1002 10:09:31.613967 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.070788458 podStartE2EDuration="7.613948986s" podCreationTimestamp="2025-10-02 10:09:24 +0000 UTC" firstStartedPulling="2025-10-02 10:09:25.684514389 +0000 UTC m=+1237.437155901" lastFinishedPulling="2025-10-02 10:09:30.227674907 +0000 UTC m=+1241.980316429" observedRunningTime="2025-10-02 10:09:31.606377048 +0000 UTC m=+1243.359018570" watchObservedRunningTime="2025-10-02 10:09:31.613948986 +0000 UTC m=+1243.366590508" Oct 02 10:09:31 crc kubenswrapper[4934]: I1002 10:09:31.634712 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.219011301 podStartE2EDuration="7.63469017s" podCreationTimestamp="2025-10-02 10:09:24 +0000 UTC" firstStartedPulling="2025-10-02 10:09:25.812004738 +0000 UTC m=+1237.564646260" lastFinishedPulling="2025-10-02 10:09:30.227683607 +0000 UTC m=+1241.980325129" observedRunningTime="2025-10-02 10:09:31.630508464 +0000 UTC m=+1243.383149996" watchObservedRunningTime="2025-10-02 10:09:31.63469017 +0000 UTC m=+1243.387331692" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.062725 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.174131 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9brt\" (UniqueName: \"kubernetes.io/projected/0d5014eb-503f-45ea-bb78-bc5978dbaa69-kube-api-access-s9brt\") pod \"0d5014eb-503f-45ea-bb78-bc5978dbaa69\" (UID: \"0d5014eb-503f-45ea-bb78-bc5978dbaa69\") " Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.181840 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d5014eb-503f-45ea-bb78-bc5978dbaa69-kube-api-access-s9brt" (OuterVolumeSpecName: "kube-api-access-s9brt") pod "0d5014eb-503f-45ea-bb78-bc5978dbaa69" (UID: "0d5014eb-503f-45ea-bb78-bc5978dbaa69"). InnerVolumeSpecName "kube-api-access-s9brt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.227787 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.275667 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cb6eb34-3995-49cb-b6a4-afeb25641e6a-combined-ca-bundle\") pod \"7cb6eb34-3995-49cb-b6a4-afeb25641e6a\" (UID: \"7cb6eb34-3995-49cb-b6a4-afeb25641e6a\") " Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.276011 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wx498\" (UniqueName: \"kubernetes.io/projected/7cb6eb34-3995-49cb-b6a4-afeb25641e6a-kube-api-access-wx498\") pod \"7cb6eb34-3995-49cb-b6a4-afeb25641e6a\" (UID: \"7cb6eb34-3995-49cb-b6a4-afeb25641e6a\") " Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.276265 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cb6eb34-3995-49cb-b6a4-afeb25641e6a-config-data\") pod \"7cb6eb34-3995-49cb-b6a4-afeb25641e6a\" (UID: \"7cb6eb34-3995-49cb-b6a4-afeb25641e6a\") " Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.276388 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cb6eb34-3995-49cb-b6a4-afeb25641e6a-logs\") pod \"7cb6eb34-3995-49cb-b6a4-afeb25641e6a\" (UID: \"7cb6eb34-3995-49cb-b6a4-afeb25641e6a\") " Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.277215 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9brt\" (UniqueName: \"kubernetes.io/projected/0d5014eb-503f-45ea-bb78-bc5978dbaa69-kube-api-access-s9brt\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.277711 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cb6eb34-3995-49cb-b6a4-afeb25641e6a-logs" (OuterVolumeSpecName: "logs") pod "7cb6eb34-3995-49cb-b6a4-afeb25641e6a" (UID: "7cb6eb34-3995-49cb-b6a4-afeb25641e6a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.282131 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cb6eb34-3995-49cb-b6a4-afeb25641e6a-kube-api-access-wx498" (OuterVolumeSpecName: "kube-api-access-wx498") pod "7cb6eb34-3995-49cb-b6a4-afeb25641e6a" (UID: "7cb6eb34-3995-49cb-b6a4-afeb25641e6a"). InnerVolumeSpecName "kube-api-access-wx498". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.316712 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cb6eb34-3995-49cb-b6a4-afeb25641e6a-config-data" (OuterVolumeSpecName: "config-data") pod "7cb6eb34-3995-49cb-b6a4-afeb25641e6a" (UID: "7cb6eb34-3995-49cb-b6a4-afeb25641e6a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.331842 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cb6eb34-3995-49cb-b6a4-afeb25641e6a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7cb6eb34-3995-49cb-b6a4-afeb25641e6a" (UID: "7cb6eb34-3995-49cb-b6a4-afeb25641e6a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.378603 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cb6eb34-3995-49cb-b6a4-afeb25641e6a-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.378637 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cb6eb34-3995-49cb-b6a4-afeb25641e6a-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.378649 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cb6eb34-3995-49cb-b6a4-afeb25641e6a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.378660 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wx498\" (UniqueName: \"kubernetes.io/projected/7cb6eb34-3995-49cb-b6a4-afeb25641e6a-kube-api-access-wx498\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.562182 4934 generic.go:334] "Generic (PLEG): container finished" podID="0d5014eb-503f-45ea-bb78-bc5978dbaa69" containerID="caccc0b41d120be902b372c1c34ce974b7a351bec8cbcf762fb7d2fb779776dd" exitCode=2 Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.562328 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"0d5014eb-503f-45ea-bb78-bc5978dbaa69","Type":"ContainerDied","Data":"caccc0b41d120be902b372c1c34ce974b7a351bec8cbcf762fb7d2fb779776dd"} Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.562537 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"0d5014eb-503f-45ea-bb78-bc5978dbaa69","Type":"ContainerDied","Data":"96cc98d752154db530e353bca1dfee39ff89568e21354a3d2231ca857d5192b0"} Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.562559 4934 scope.go:117] "RemoveContainer" containerID="caccc0b41d120be902b372c1c34ce974b7a351bec8cbcf762fb7d2fb779776dd" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.562412 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.568078 4934 generic.go:334] "Generic (PLEG): container finished" podID="7cb6eb34-3995-49cb-b6a4-afeb25641e6a" containerID="7f84fb7c6e9e5d88dc364e7a4dfa9e56d0e2b98f171280f1351177ac7b560759" exitCode=0 Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.568108 4934 generic.go:334] "Generic (PLEG): container finished" podID="7cb6eb34-3995-49cb-b6a4-afeb25641e6a" containerID="b077cf2ec73538e90e92ac14e7c3ac51e222f05cddaeb92375d89df43e635b1e" exitCode=143 Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.568150 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.568216 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7cb6eb34-3995-49cb-b6a4-afeb25641e6a","Type":"ContainerDied","Data":"7f84fb7c6e9e5d88dc364e7a4dfa9e56d0e2b98f171280f1351177ac7b560759"} Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.568282 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7cb6eb34-3995-49cb-b6a4-afeb25641e6a","Type":"ContainerDied","Data":"b077cf2ec73538e90e92ac14e7c3ac51e222f05cddaeb92375d89df43e635b1e"} Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.568296 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7cb6eb34-3995-49cb-b6a4-afeb25641e6a","Type":"ContainerDied","Data":"dd7866576be4b5cf7f5de0102b5c7667c430898b16a4aa2b636e0b581019d815"} Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.592071 4934 scope.go:117] "RemoveContainer" containerID="caccc0b41d120be902b372c1c34ce974b7a351bec8cbcf762fb7d2fb779776dd" Oct 02 10:09:32 crc kubenswrapper[4934]: E1002 10:09:32.592669 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"caccc0b41d120be902b372c1c34ce974b7a351bec8cbcf762fb7d2fb779776dd\": container with ID starting with caccc0b41d120be902b372c1c34ce974b7a351bec8cbcf762fb7d2fb779776dd not found: ID does not exist" containerID="caccc0b41d120be902b372c1c34ce974b7a351bec8cbcf762fb7d2fb779776dd" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.592708 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"caccc0b41d120be902b372c1c34ce974b7a351bec8cbcf762fb7d2fb779776dd"} err="failed to get container status \"caccc0b41d120be902b372c1c34ce974b7a351bec8cbcf762fb7d2fb779776dd\": rpc error: code = NotFound desc = could not find container \"caccc0b41d120be902b372c1c34ce974b7a351bec8cbcf762fb7d2fb779776dd\": container with ID starting with caccc0b41d120be902b372c1c34ce974b7a351bec8cbcf762fb7d2fb779776dd not found: ID does not exist" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.592736 4934 scope.go:117] "RemoveContainer" containerID="7f84fb7c6e9e5d88dc364e7a4dfa9e56d0e2b98f171280f1351177ac7b560759" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.602175 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.618645 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.630477 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.634773 4934 scope.go:117] "RemoveContainer" containerID="b077cf2ec73538e90e92ac14e7c3ac51e222f05cddaeb92375d89df43e635b1e" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.649206 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.660646 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 10:09:32 crc kubenswrapper[4934]: E1002 10:09:32.661296 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cb6eb34-3995-49cb-b6a4-afeb25641e6a" containerName="nova-metadata-log" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.661393 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cb6eb34-3995-49cb-b6a4-afeb25641e6a" containerName="nova-metadata-log" Oct 02 10:09:32 crc kubenswrapper[4934]: E1002 10:09:32.661472 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d5014eb-503f-45ea-bb78-bc5978dbaa69" containerName="kube-state-metrics" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.661528 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d5014eb-503f-45ea-bb78-bc5978dbaa69" containerName="kube-state-metrics" Oct 02 10:09:32 crc kubenswrapper[4934]: E1002 10:09:32.661643 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cb6eb34-3995-49cb-b6a4-afeb25641e6a" containerName="nova-metadata-metadata" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.661735 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cb6eb34-3995-49cb-b6a4-afeb25641e6a" containerName="nova-metadata-metadata" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.661987 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d5014eb-503f-45ea-bb78-bc5978dbaa69" containerName="kube-state-metrics" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.662069 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cb6eb34-3995-49cb-b6a4-afeb25641e6a" containerName="nova-metadata-log" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.662139 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cb6eb34-3995-49cb-b6a4-afeb25641e6a" containerName="nova-metadata-metadata" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.662870 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.667915 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.668091 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.670064 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.688930 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np29c\" (UniqueName: \"kubernetes.io/projected/958bc92f-fd6d-4ba0-a21f-8f1c58bac789-kube-api-access-np29c\") pod \"kube-state-metrics-0\" (UID: \"958bc92f-fd6d-4ba0-a21f-8f1c58bac789\") " pod="openstack/kube-state-metrics-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.689048 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/958bc92f-fd6d-4ba0-a21f-8f1c58bac789-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"958bc92f-fd6d-4ba0-a21f-8f1c58bac789\") " pod="openstack/kube-state-metrics-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.689171 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/958bc92f-fd6d-4ba0-a21f-8f1c58bac789-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"958bc92f-fd6d-4ba0-a21f-8f1c58bac789\") " pod="openstack/kube-state-metrics-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.689225 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/958bc92f-fd6d-4ba0-a21f-8f1c58bac789-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"958bc92f-fd6d-4ba0-a21f-8f1c58bac789\") " pod="openstack/kube-state-metrics-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.691803 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.696397 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.700145 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.700191 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.705651 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.711021 4934 scope.go:117] "RemoveContainer" containerID="7f84fb7c6e9e5d88dc364e7a4dfa9e56d0e2b98f171280f1351177ac7b560759" Oct 02 10:09:32 crc kubenswrapper[4934]: E1002 10:09:32.714708 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f84fb7c6e9e5d88dc364e7a4dfa9e56d0e2b98f171280f1351177ac7b560759\": container with ID starting with 7f84fb7c6e9e5d88dc364e7a4dfa9e56d0e2b98f171280f1351177ac7b560759 not found: ID does not exist" containerID="7f84fb7c6e9e5d88dc364e7a4dfa9e56d0e2b98f171280f1351177ac7b560759" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.714774 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f84fb7c6e9e5d88dc364e7a4dfa9e56d0e2b98f171280f1351177ac7b560759"} err="failed to get container status \"7f84fb7c6e9e5d88dc364e7a4dfa9e56d0e2b98f171280f1351177ac7b560759\": rpc error: code = NotFound desc = could not find container \"7f84fb7c6e9e5d88dc364e7a4dfa9e56d0e2b98f171280f1351177ac7b560759\": container with ID starting with 7f84fb7c6e9e5d88dc364e7a4dfa9e56d0e2b98f171280f1351177ac7b560759 not found: ID does not exist" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.714804 4934 scope.go:117] "RemoveContainer" containerID="b077cf2ec73538e90e92ac14e7c3ac51e222f05cddaeb92375d89df43e635b1e" Oct 02 10:09:32 crc kubenswrapper[4934]: E1002 10:09:32.716878 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b077cf2ec73538e90e92ac14e7c3ac51e222f05cddaeb92375d89df43e635b1e\": container with ID starting with b077cf2ec73538e90e92ac14e7c3ac51e222f05cddaeb92375d89df43e635b1e not found: ID does not exist" containerID="b077cf2ec73538e90e92ac14e7c3ac51e222f05cddaeb92375d89df43e635b1e" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.716908 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b077cf2ec73538e90e92ac14e7c3ac51e222f05cddaeb92375d89df43e635b1e"} err="failed to get container status \"b077cf2ec73538e90e92ac14e7c3ac51e222f05cddaeb92375d89df43e635b1e\": rpc error: code = NotFound desc = could not find container \"b077cf2ec73538e90e92ac14e7c3ac51e222f05cddaeb92375d89df43e635b1e\": container with ID starting with b077cf2ec73538e90e92ac14e7c3ac51e222f05cddaeb92375d89df43e635b1e not found: ID does not exist" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.716928 4934 scope.go:117] "RemoveContainer" containerID="7f84fb7c6e9e5d88dc364e7a4dfa9e56d0e2b98f171280f1351177ac7b560759" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.717731 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f84fb7c6e9e5d88dc364e7a4dfa9e56d0e2b98f171280f1351177ac7b560759"} err="failed to get container status \"7f84fb7c6e9e5d88dc364e7a4dfa9e56d0e2b98f171280f1351177ac7b560759\": rpc error: code = NotFound desc = could not find container \"7f84fb7c6e9e5d88dc364e7a4dfa9e56d0e2b98f171280f1351177ac7b560759\": container with ID starting with 7f84fb7c6e9e5d88dc364e7a4dfa9e56d0e2b98f171280f1351177ac7b560759 not found: ID does not exist" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.717776 4934 scope.go:117] "RemoveContainer" containerID="b077cf2ec73538e90e92ac14e7c3ac51e222f05cddaeb92375d89df43e635b1e" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.719839 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b077cf2ec73538e90e92ac14e7c3ac51e222f05cddaeb92375d89df43e635b1e"} err="failed to get container status \"b077cf2ec73538e90e92ac14e7c3ac51e222f05cddaeb92375d89df43e635b1e\": rpc error: code = NotFound desc = could not find container \"b077cf2ec73538e90e92ac14e7c3ac51e222f05cddaeb92375d89df43e635b1e\": container with ID starting with b077cf2ec73538e90e92ac14e7c3ac51e222f05cddaeb92375d89df43e635b1e not found: ID does not exist" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.792334 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/958bc92f-fd6d-4ba0-a21f-8f1c58bac789-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"958bc92f-fd6d-4ba0-a21f-8f1c58bac789\") " pod="openstack/kube-state-metrics-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.792419 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/958bc92f-fd6d-4ba0-a21f-8f1c58bac789-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"958bc92f-fd6d-4ba0-a21f-8f1c58bac789\") " pod="openstack/kube-state-metrics-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.792472 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17364b5a-18b8-4b54-a675-ca2c3b0d110b-logs\") pod \"nova-metadata-0\" (UID: \"17364b5a-18b8-4b54-a675-ca2c3b0d110b\") " pod="openstack/nova-metadata-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.792508 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17364b5a-18b8-4b54-a675-ca2c3b0d110b-config-data\") pod \"nova-metadata-0\" (UID: \"17364b5a-18b8-4b54-a675-ca2c3b0d110b\") " pod="openstack/nova-metadata-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.792563 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/17364b5a-18b8-4b54-a675-ca2c3b0d110b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"17364b5a-18b8-4b54-a675-ca2c3b0d110b\") " pod="openstack/nova-metadata-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.792660 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7v56m\" (UniqueName: \"kubernetes.io/projected/17364b5a-18b8-4b54-a675-ca2c3b0d110b-kube-api-access-7v56m\") pod \"nova-metadata-0\" (UID: \"17364b5a-18b8-4b54-a675-ca2c3b0d110b\") " pod="openstack/nova-metadata-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.792709 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np29c\" (UniqueName: \"kubernetes.io/projected/958bc92f-fd6d-4ba0-a21f-8f1c58bac789-kube-api-access-np29c\") pod \"kube-state-metrics-0\" (UID: \"958bc92f-fd6d-4ba0-a21f-8f1c58bac789\") " pod="openstack/kube-state-metrics-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.792738 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17364b5a-18b8-4b54-a675-ca2c3b0d110b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"17364b5a-18b8-4b54-a675-ca2c3b0d110b\") " pod="openstack/nova-metadata-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.792884 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/958bc92f-fd6d-4ba0-a21f-8f1c58bac789-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"958bc92f-fd6d-4ba0-a21f-8f1c58bac789\") " pod="openstack/kube-state-metrics-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.798178 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/958bc92f-fd6d-4ba0-a21f-8f1c58bac789-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"958bc92f-fd6d-4ba0-a21f-8f1c58bac789\") " pod="openstack/kube-state-metrics-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.811176 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/958bc92f-fd6d-4ba0-a21f-8f1c58bac789-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"958bc92f-fd6d-4ba0-a21f-8f1c58bac789\") " pod="openstack/kube-state-metrics-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.814153 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/958bc92f-fd6d-4ba0-a21f-8f1c58bac789-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"958bc92f-fd6d-4ba0-a21f-8f1c58bac789\") " pod="openstack/kube-state-metrics-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.816204 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np29c\" (UniqueName: \"kubernetes.io/projected/958bc92f-fd6d-4ba0-a21f-8f1c58bac789-kube-api-access-np29c\") pod \"kube-state-metrics-0\" (UID: \"958bc92f-fd6d-4ba0-a21f-8f1c58bac789\") " pod="openstack/kube-state-metrics-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.894348 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17364b5a-18b8-4b54-a675-ca2c3b0d110b-logs\") pod \"nova-metadata-0\" (UID: \"17364b5a-18b8-4b54-a675-ca2c3b0d110b\") " pod="openstack/nova-metadata-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.894398 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17364b5a-18b8-4b54-a675-ca2c3b0d110b-config-data\") pod \"nova-metadata-0\" (UID: \"17364b5a-18b8-4b54-a675-ca2c3b0d110b\") " pod="openstack/nova-metadata-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.894421 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/17364b5a-18b8-4b54-a675-ca2c3b0d110b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"17364b5a-18b8-4b54-a675-ca2c3b0d110b\") " pod="openstack/nova-metadata-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.894450 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7v56m\" (UniqueName: \"kubernetes.io/projected/17364b5a-18b8-4b54-a675-ca2c3b0d110b-kube-api-access-7v56m\") pod \"nova-metadata-0\" (UID: \"17364b5a-18b8-4b54-a675-ca2c3b0d110b\") " pod="openstack/nova-metadata-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.894484 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17364b5a-18b8-4b54-a675-ca2c3b0d110b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"17364b5a-18b8-4b54-a675-ca2c3b0d110b\") " pod="openstack/nova-metadata-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.896044 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17364b5a-18b8-4b54-a675-ca2c3b0d110b-logs\") pod \"nova-metadata-0\" (UID: \"17364b5a-18b8-4b54-a675-ca2c3b0d110b\") " pod="openstack/nova-metadata-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.900012 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/17364b5a-18b8-4b54-a675-ca2c3b0d110b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"17364b5a-18b8-4b54-a675-ca2c3b0d110b\") " pod="openstack/nova-metadata-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.900213 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17364b5a-18b8-4b54-a675-ca2c3b0d110b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"17364b5a-18b8-4b54-a675-ca2c3b0d110b\") " pod="openstack/nova-metadata-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.900385 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17364b5a-18b8-4b54-a675-ca2c3b0d110b-config-data\") pod \"nova-metadata-0\" (UID: \"17364b5a-18b8-4b54-a675-ca2c3b0d110b\") " pod="openstack/nova-metadata-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.910154 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7v56m\" (UniqueName: \"kubernetes.io/projected/17364b5a-18b8-4b54-a675-ca2c3b0d110b-kube-api-access-7v56m\") pod \"nova-metadata-0\" (UID: \"17364b5a-18b8-4b54-a675-ca2c3b0d110b\") " pod="openstack/nova-metadata-0" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.923536 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d5014eb-503f-45ea-bb78-bc5978dbaa69" path="/var/lib/kubelet/pods/0d5014eb-503f-45ea-bb78-bc5978dbaa69/volumes" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.924154 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cb6eb34-3995-49cb-b6a4-afeb25641e6a" path="/var/lib/kubelet/pods/7cb6eb34-3995-49cb-b6a4-afeb25641e6a/volumes" Oct 02 10:09:32 crc kubenswrapper[4934]: I1002 10:09:32.991329 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 10:09:33 crc kubenswrapper[4934]: I1002 10:09:33.027292 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 10:09:33 crc kubenswrapper[4934]: I1002 10:09:33.458242 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 10:09:33 crc kubenswrapper[4934]: I1002 10:09:33.582530 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"958bc92f-fd6d-4ba0-a21f-8f1c58bac789","Type":"ContainerStarted","Data":"6653b74890db6090b03cbdefd5f76446661ab9976bd658ec66207c6f4414e1e8"} Oct 02 10:09:33 crc kubenswrapper[4934]: I1002 10:09:33.610784 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:09:33 crc kubenswrapper[4934]: I1002 10:09:33.762515 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:09:33 crc kubenswrapper[4934]: I1002 10:09:33.762820 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="06eb14a8-953f-48f9-9da1-a217d6f5568a" containerName="ceilometer-central-agent" containerID="cri-o://c630cb405e8c2e671298d31e62bc1dd4507dff91d4e9081ba2703b0a2c966f1a" gracePeriod=30 Oct 02 10:09:33 crc kubenswrapper[4934]: I1002 10:09:33.762860 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="06eb14a8-953f-48f9-9da1-a217d6f5568a" containerName="ceilometer-notification-agent" containerID="cri-o://3a12e64befe2bba1ea2f15bb3318dd9358c70985e22da881b9157cfb0998d0a1" gracePeriod=30 Oct 02 10:09:33 crc kubenswrapper[4934]: I1002 10:09:33.762825 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="06eb14a8-953f-48f9-9da1-a217d6f5568a" containerName="proxy-httpd" containerID="cri-o://1b6f098eddfd42439826b80bb65f1c06a66432a5934703606b4954bb0e943de3" gracePeriod=30 Oct 02 10:09:33 crc kubenswrapper[4934]: I1002 10:09:33.762886 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="06eb14a8-953f-48f9-9da1-a217d6f5568a" containerName="sg-core" containerID="cri-o://1e06b21ceb0c24b7ae0cae51c6becaf3172bbda2e4d646bf490f1cc858852347" gracePeriod=30 Oct 02 10:09:34 crc kubenswrapper[4934]: I1002 10:09:34.597495 4934 generic.go:334] "Generic (PLEG): container finished" podID="ed820a81-72fc-4d84-876f-9def630d4ff7" containerID="8d3c9f48dcfcd16e834077a1b8cefb91be17eaec6d3b75578d4108c226272f38" exitCode=0 Oct 02 10:09:34 crc kubenswrapper[4934]: I1002 10:09:34.597774 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zkqqc" event={"ID":"ed820a81-72fc-4d84-876f-9def630d4ff7","Type":"ContainerDied","Data":"8d3c9f48dcfcd16e834077a1b8cefb91be17eaec6d3b75578d4108c226272f38"} Oct 02 10:09:34 crc kubenswrapper[4934]: I1002 10:09:34.610351 4934 generic.go:334] "Generic (PLEG): container finished" podID="06eb14a8-953f-48f9-9da1-a217d6f5568a" containerID="1b6f098eddfd42439826b80bb65f1c06a66432a5934703606b4954bb0e943de3" exitCode=0 Oct 02 10:09:34 crc kubenswrapper[4934]: I1002 10:09:34.610386 4934 generic.go:334] "Generic (PLEG): container finished" podID="06eb14a8-953f-48f9-9da1-a217d6f5568a" containerID="1e06b21ceb0c24b7ae0cae51c6becaf3172bbda2e4d646bf490f1cc858852347" exitCode=2 Oct 02 10:09:34 crc kubenswrapper[4934]: I1002 10:09:34.610394 4934 generic.go:334] "Generic (PLEG): container finished" podID="06eb14a8-953f-48f9-9da1-a217d6f5568a" containerID="c630cb405e8c2e671298d31e62bc1dd4507dff91d4e9081ba2703b0a2c966f1a" exitCode=0 Oct 02 10:09:34 crc kubenswrapper[4934]: I1002 10:09:34.610455 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06eb14a8-953f-48f9-9da1-a217d6f5568a","Type":"ContainerDied","Data":"1b6f098eddfd42439826b80bb65f1c06a66432a5934703606b4954bb0e943de3"} Oct 02 10:09:34 crc kubenswrapper[4934]: I1002 10:09:34.610483 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06eb14a8-953f-48f9-9da1-a217d6f5568a","Type":"ContainerDied","Data":"1e06b21ceb0c24b7ae0cae51c6becaf3172bbda2e4d646bf490f1cc858852347"} Oct 02 10:09:34 crc kubenswrapper[4934]: I1002 10:09:34.610493 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06eb14a8-953f-48f9-9da1-a217d6f5568a","Type":"ContainerDied","Data":"c630cb405e8c2e671298d31e62bc1dd4507dff91d4e9081ba2703b0a2c966f1a"} Oct 02 10:09:34 crc kubenswrapper[4934]: I1002 10:09:34.628452 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"17364b5a-18b8-4b54-a675-ca2c3b0d110b","Type":"ContainerStarted","Data":"64d5c9337cd70862c04d06eea9b926f4290509b8e0b09d73de7a3ea0c181811e"} Oct 02 10:09:34 crc kubenswrapper[4934]: I1002 10:09:34.628492 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"17364b5a-18b8-4b54-a675-ca2c3b0d110b","Type":"ContainerStarted","Data":"af70f69dc380a109c459550553d05155d01a8a3bbf85e90bbafd0e406fa3c047"} Oct 02 10:09:34 crc kubenswrapper[4934]: I1002 10:09:34.628501 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"17364b5a-18b8-4b54-a675-ca2c3b0d110b","Type":"ContainerStarted","Data":"c8dfb9c8c731e7e7f765eff24745efbaf2f4015ebfed63f3824811d8d1d5ed00"} Oct 02 10:09:34 crc kubenswrapper[4934]: I1002 10:09:34.634102 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"958bc92f-fd6d-4ba0-a21f-8f1c58bac789","Type":"ContainerStarted","Data":"877a883b01af6709bd26ee32e0a326b213d51f06b683c07ca0c300ae16ad52db"} Oct 02 10:09:34 crc kubenswrapper[4934]: I1002 10:09:34.634817 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 02 10:09:34 crc kubenswrapper[4934]: I1002 10:09:34.687871 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.088353625 podStartE2EDuration="2.687852175s" podCreationTimestamp="2025-10-02 10:09:32 +0000 UTC" firstStartedPulling="2025-10-02 10:09:33.471484025 +0000 UTC m=+1245.224125547" lastFinishedPulling="2025-10-02 10:09:34.070982575 +0000 UTC m=+1245.823624097" observedRunningTime="2025-10-02 10:09:34.682694322 +0000 UTC m=+1246.435335864" watchObservedRunningTime="2025-10-02 10:09:34.687852175 +0000 UTC m=+1246.440493697" Oct 02 10:09:34 crc kubenswrapper[4934]: I1002 10:09:34.702388 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.702366025 podStartE2EDuration="2.702366025s" podCreationTimestamp="2025-10-02 10:09:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:09:34.655315466 +0000 UTC m=+1246.407956988" watchObservedRunningTime="2025-10-02 10:09:34.702366025 +0000 UTC m=+1246.455007547" Oct 02 10:09:34 crc kubenswrapper[4934]: I1002 10:09:34.851324 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:09:35 crc kubenswrapper[4934]: I1002 10:09:35.175589 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 10:09:35 crc kubenswrapper[4934]: I1002 10:09:35.176037 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 10:09:35 crc kubenswrapper[4934]: I1002 10:09:35.213342 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 10:09:35 crc kubenswrapper[4934]: I1002 10:09:35.249963 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 10:09:35 crc kubenswrapper[4934]: I1002 10:09:35.250012 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 10:09:35 crc kubenswrapper[4934]: I1002 10:09:35.269723 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" Oct 02 10:09:35 crc kubenswrapper[4934]: I1002 10:09:35.331172 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55d67bcc45-h5d2q"] Oct 02 10:09:35 crc kubenswrapper[4934]: I1002 10:09:35.331565 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" podUID="73e6e086-8443-46e5-89f8-21bc9bde6b4f" containerName="dnsmasq-dns" containerID="cri-o://867e46f0e9e66fc2f4cda4cf5adfa64cd1fa5ad8df0d05ee0a276ebbba2ce317" gracePeriod=10 Oct 02 10:09:35 crc kubenswrapper[4934]: I1002 10:09:35.655964 4934 generic.go:334] "Generic (PLEG): container finished" podID="73e6e086-8443-46e5-89f8-21bc9bde6b4f" containerID="867e46f0e9e66fc2f4cda4cf5adfa64cd1fa5ad8df0d05ee0a276ebbba2ce317" exitCode=0 Oct 02 10:09:35 crc kubenswrapper[4934]: I1002 10:09:35.656235 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" event={"ID":"73e6e086-8443-46e5-89f8-21bc9bde6b4f","Type":"ContainerDied","Data":"867e46f0e9e66fc2f4cda4cf5adfa64cd1fa5ad8df0d05ee0a276ebbba2ce317"} Oct 02 10:09:35 crc kubenswrapper[4934]: I1002 10:09:35.701771 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 10:09:35 crc kubenswrapper[4934]: I1002 10:09:35.872388 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" Oct 02 10:09:35 crc kubenswrapper[4934]: I1002 10:09:35.958162 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-config\") pod \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\" (UID: \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\") " Oct 02 10:09:35 crc kubenswrapper[4934]: I1002 10:09:35.958203 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-ovsdbserver-nb\") pod \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\" (UID: \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\") " Oct 02 10:09:35 crc kubenswrapper[4934]: I1002 10:09:35.958250 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2bln\" (UniqueName: \"kubernetes.io/projected/73e6e086-8443-46e5-89f8-21bc9bde6b4f-kube-api-access-g2bln\") pod \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\" (UID: \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\") " Oct 02 10:09:35 crc kubenswrapper[4934]: I1002 10:09:35.958269 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-ovsdbserver-sb\") pod \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\" (UID: \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\") " Oct 02 10:09:35 crc kubenswrapper[4934]: I1002 10:09:35.958408 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-dns-svc\") pod \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\" (UID: \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\") " Oct 02 10:09:35 crc kubenswrapper[4934]: I1002 10:09:35.958426 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-dns-swift-storage-0\") pod \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\" (UID: \"73e6e086-8443-46e5-89f8-21bc9bde6b4f\") " Oct 02 10:09:35 crc kubenswrapper[4934]: I1002 10:09:35.971689 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73e6e086-8443-46e5-89f8-21bc9bde6b4f-kube-api-access-g2bln" (OuterVolumeSpecName: "kube-api-access-g2bln") pod "73e6e086-8443-46e5-89f8-21bc9bde6b4f" (UID: "73e6e086-8443-46e5-89f8-21bc9bde6b4f"). InnerVolumeSpecName "kube-api-access-g2bln". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.057494 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "73e6e086-8443-46e5-89f8-21bc9bde6b4f" (UID: "73e6e086-8443-46e5-89f8-21bc9bde6b4f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.061273 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2bln\" (UniqueName: \"kubernetes.io/projected/73e6e086-8443-46e5-89f8-21bc9bde6b4f-kube-api-access-g2bln\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.061301 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.071056 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "73e6e086-8443-46e5-89f8-21bc9bde6b4f" (UID: "73e6e086-8443-46e5-89f8-21bc9bde6b4f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.073971 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-config" (OuterVolumeSpecName: "config") pod "73e6e086-8443-46e5-89f8-21bc9bde6b4f" (UID: "73e6e086-8443-46e5-89f8-21bc9bde6b4f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.076467 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "73e6e086-8443-46e5-89f8-21bc9bde6b4f" (UID: "73e6e086-8443-46e5-89f8-21bc9bde6b4f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.082059 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "73e6e086-8443-46e5-89f8-21bc9bde6b4f" (UID: "73e6e086-8443-46e5-89f8-21bc9bde6b4f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.124126 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zkqqc" Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.171230 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed820a81-72fc-4d84-876f-9def630d4ff7-combined-ca-bundle\") pod \"ed820a81-72fc-4d84-876f-9def630d4ff7\" (UID: \"ed820a81-72fc-4d84-876f-9def630d4ff7\") " Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.171331 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed820a81-72fc-4d84-876f-9def630d4ff7-config-data\") pod \"ed820a81-72fc-4d84-876f-9def630d4ff7\" (UID: \"ed820a81-72fc-4d84-876f-9def630d4ff7\") " Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.171371 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed820a81-72fc-4d84-876f-9def630d4ff7-scripts\") pod \"ed820a81-72fc-4d84-876f-9def630d4ff7\" (UID: \"ed820a81-72fc-4d84-876f-9def630d4ff7\") " Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.171514 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xj9ff\" (UniqueName: \"kubernetes.io/projected/ed820a81-72fc-4d84-876f-9def630d4ff7-kube-api-access-xj9ff\") pod \"ed820a81-72fc-4d84-876f-9def630d4ff7\" (UID: \"ed820a81-72fc-4d84-876f-9def630d4ff7\") " Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.171970 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.171987 4934 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.171997 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.172005 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73e6e086-8443-46e5-89f8-21bc9bde6b4f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.177037 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed820a81-72fc-4d84-876f-9def630d4ff7-scripts" (OuterVolumeSpecName: "scripts") pod "ed820a81-72fc-4d84-876f-9def630d4ff7" (UID: "ed820a81-72fc-4d84-876f-9def630d4ff7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.180973 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed820a81-72fc-4d84-876f-9def630d4ff7-kube-api-access-xj9ff" (OuterVolumeSpecName: "kube-api-access-xj9ff") pod "ed820a81-72fc-4d84-876f-9def630d4ff7" (UID: "ed820a81-72fc-4d84-876f-9def630d4ff7"). InnerVolumeSpecName "kube-api-access-xj9ff". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.204721 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed820a81-72fc-4d84-876f-9def630d4ff7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed820a81-72fc-4d84-876f-9def630d4ff7" (UID: "ed820a81-72fc-4d84-876f-9def630d4ff7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.217925 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed820a81-72fc-4d84-876f-9def630d4ff7-config-data" (OuterVolumeSpecName: "config-data") pod "ed820a81-72fc-4d84-876f-9def630d4ff7" (UID: "ed820a81-72fc-4d84-876f-9def630d4ff7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.274064 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xj9ff\" (UniqueName: \"kubernetes.io/projected/ed820a81-72fc-4d84-876f-9def630d4ff7-kube-api-access-xj9ff\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.274343 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed820a81-72fc-4d84-876f-9def630d4ff7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.274428 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed820a81-72fc-4d84-876f-9def630d4ff7-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.274498 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed820a81-72fc-4d84-876f-9def630d4ff7-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.332748 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b4da04f7-1f59-4acc-90e0-7b7cba140393" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.188:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.332780 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b4da04f7-1f59-4acc-90e0-7b7cba140393" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.188:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.665389 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" event={"ID":"73e6e086-8443-46e5-89f8-21bc9bde6b4f","Type":"ContainerDied","Data":"e868d7e6f05fe721e79b35bdd140cbdd9d0ae1b99d3357190404fab44cc1190e"} Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.665422 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55d67bcc45-h5d2q" Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.665649 4934 scope.go:117] "RemoveContainer" containerID="867e46f0e9e66fc2f4cda4cf5adfa64cd1fa5ad8df0d05ee0a276ebbba2ce317" Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.667326 4934 generic.go:334] "Generic (PLEG): container finished" podID="490173ef-a363-4b84-8802-6a69bb9bdb76" containerID="6b77fc0e555402c151a111643d602de419a5a7a8fe3ab9b1c50acb80d1af507a" exitCode=0 Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.667391 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-6c8h9" event={"ID":"490173ef-a363-4b84-8802-6a69bb9bdb76","Type":"ContainerDied","Data":"6b77fc0e555402c151a111643d602de419a5a7a8fe3ab9b1c50acb80d1af507a"} Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.669410 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zkqqc" event={"ID":"ed820a81-72fc-4d84-876f-9def630d4ff7","Type":"ContainerDied","Data":"a1c339ab77a926bc8a5fc8a29f16b23b5fc3fe605689c09863b2cd3f4e9142c5"} Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.669587 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zkqqc" Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.669627 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1c339ab77a926bc8a5fc8a29f16b23b5fc3fe605689c09863b2cd3f4e9142c5" Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.688094 4934 scope.go:117] "RemoveContainer" containerID="685b0edf8574c2f1879850eeb6fc75939e3235e6d1e40e86f77b4f04d51d5f5e" Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.708031 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55d67bcc45-h5d2q"] Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.714924 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55d67bcc45-h5d2q"] Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.787323 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.801069 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.801283 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b4da04f7-1f59-4acc-90e0-7b7cba140393" containerName="nova-api-log" containerID="cri-o://2b6fe7b50c803ac8ece2433dd06ec23dcfef9b8a9b399c435d0bd26a0fc0e9f2" gracePeriod=30 Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.801419 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b4da04f7-1f59-4acc-90e0-7b7cba140393" containerName="nova-api-api" containerID="cri-o://a6d1113b52ffb7001a8b2d59c298de5526ba2211698784ccbcdf6bb49e0e2e01" gracePeriod=30 Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.830213 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.830536 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="17364b5a-18b8-4b54-a675-ca2c3b0d110b" containerName="nova-metadata-metadata" containerID="cri-o://64d5c9337cd70862c04d06eea9b926f4290509b8e0b09d73de7a3ea0c181811e" gracePeriod=30 Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.830692 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="17364b5a-18b8-4b54-a675-ca2c3b0d110b" containerName="nova-metadata-log" containerID="cri-o://af70f69dc380a109c459550553d05155d01a8a3bbf85e90bbafd0e406fa3c047" gracePeriod=30 Oct 02 10:09:36 crc kubenswrapper[4934]: I1002 10:09:36.928707 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73e6e086-8443-46e5-89f8-21bc9bde6b4f" path="/var/lib/kubelet/pods/73e6e086-8443-46e5-89f8-21bc9bde6b4f/volumes" Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.583789 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.596770 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7v56m\" (UniqueName: \"kubernetes.io/projected/17364b5a-18b8-4b54-a675-ca2c3b0d110b-kube-api-access-7v56m\") pod \"17364b5a-18b8-4b54-a675-ca2c3b0d110b\" (UID: \"17364b5a-18b8-4b54-a675-ca2c3b0d110b\") " Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.596863 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17364b5a-18b8-4b54-a675-ca2c3b0d110b-logs\") pod \"17364b5a-18b8-4b54-a675-ca2c3b0d110b\" (UID: \"17364b5a-18b8-4b54-a675-ca2c3b0d110b\") " Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.596988 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17364b5a-18b8-4b54-a675-ca2c3b0d110b-combined-ca-bundle\") pod \"17364b5a-18b8-4b54-a675-ca2c3b0d110b\" (UID: \"17364b5a-18b8-4b54-a675-ca2c3b0d110b\") " Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.597038 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17364b5a-18b8-4b54-a675-ca2c3b0d110b-config-data\") pod \"17364b5a-18b8-4b54-a675-ca2c3b0d110b\" (UID: \"17364b5a-18b8-4b54-a675-ca2c3b0d110b\") " Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.597120 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/17364b5a-18b8-4b54-a675-ca2c3b0d110b-nova-metadata-tls-certs\") pod \"17364b5a-18b8-4b54-a675-ca2c3b0d110b\" (UID: \"17364b5a-18b8-4b54-a675-ca2c3b0d110b\") " Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.597328 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17364b5a-18b8-4b54-a675-ca2c3b0d110b-logs" (OuterVolumeSpecName: "logs") pod "17364b5a-18b8-4b54-a675-ca2c3b0d110b" (UID: "17364b5a-18b8-4b54-a675-ca2c3b0d110b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.597893 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17364b5a-18b8-4b54-a675-ca2c3b0d110b-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.606838 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17364b5a-18b8-4b54-a675-ca2c3b0d110b-kube-api-access-7v56m" (OuterVolumeSpecName: "kube-api-access-7v56m") pod "17364b5a-18b8-4b54-a675-ca2c3b0d110b" (UID: "17364b5a-18b8-4b54-a675-ca2c3b0d110b"). InnerVolumeSpecName "kube-api-access-7v56m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.639655 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17364b5a-18b8-4b54-a675-ca2c3b0d110b-config-data" (OuterVolumeSpecName: "config-data") pod "17364b5a-18b8-4b54-a675-ca2c3b0d110b" (UID: "17364b5a-18b8-4b54-a675-ca2c3b0d110b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.639756 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17364b5a-18b8-4b54-a675-ca2c3b0d110b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "17364b5a-18b8-4b54-a675-ca2c3b0d110b" (UID: "17364b5a-18b8-4b54-a675-ca2c3b0d110b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.699665 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7v56m\" (UniqueName: \"kubernetes.io/projected/17364b5a-18b8-4b54-a675-ca2c3b0d110b-kube-api-access-7v56m\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.699704 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17364b5a-18b8-4b54-a675-ca2c3b0d110b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.699717 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17364b5a-18b8-4b54-a675-ca2c3b0d110b-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.704321 4934 generic.go:334] "Generic (PLEG): container finished" podID="b4da04f7-1f59-4acc-90e0-7b7cba140393" containerID="2b6fe7b50c803ac8ece2433dd06ec23dcfef9b8a9b399c435d0bd26a0fc0e9f2" exitCode=143 Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.704426 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b4da04f7-1f59-4acc-90e0-7b7cba140393","Type":"ContainerDied","Data":"2b6fe7b50c803ac8ece2433dd06ec23dcfef9b8a9b399c435d0bd26a0fc0e9f2"} Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.712234 4934 generic.go:334] "Generic (PLEG): container finished" podID="17364b5a-18b8-4b54-a675-ca2c3b0d110b" containerID="64d5c9337cd70862c04d06eea9b926f4290509b8e0b09d73de7a3ea0c181811e" exitCode=0 Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.712272 4934 generic.go:334] "Generic (PLEG): container finished" podID="17364b5a-18b8-4b54-a675-ca2c3b0d110b" containerID="af70f69dc380a109c459550553d05155d01a8a3bbf85e90bbafd0e406fa3c047" exitCode=143 Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.712469 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.712352 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"17364b5a-18b8-4b54-a675-ca2c3b0d110b","Type":"ContainerDied","Data":"64d5c9337cd70862c04d06eea9b926f4290509b8e0b09d73de7a3ea0c181811e"} Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.712613 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"17364b5a-18b8-4b54-a675-ca2c3b0d110b","Type":"ContainerDied","Data":"af70f69dc380a109c459550553d05155d01a8a3bbf85e90bbafd0e406fa3c047"} Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.712652 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"17364b5a-18b8-4b54-a675-ca2c3b0d110b","Type":"ContainerDied","Data":"c8dfb9c8c731e7e7f765eff24745efbaf2f4015ebfed63f3824811d8d1d5ed00"} Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.712675 4934 scope.go:117] "RemoveContainer" containerID="64d5c9337cd70862c04d06eea9b926f4290509b8e0b09d73de7a3ea0c181811e" Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.716202 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17364b5a-18b8-4b54-a675-ca2c3b0d110b-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "17364b5a-18b8-4b54-a675-ca2c3b0d110b" (UID: "17364b5a-18b8-4b54-a675-ca2c3b0d110b"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.756568 4934 scope.go:117] "RemoveContainer" containerID="af70f69dc380a109c459550553d05155d01a8a3bbf85e90bbafd0e406fa3c047" Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.778752 4934 scope.go:117] "RemoveContainer" containerID="64d5c9337cd70862c04d06eea9b926f4290509b8e0b09d73de7a3ea0c181811e" Oct 02 10:09:37 crc kubenswrapper[4934]: E1002 10:09:37.779152 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64d5c9337cd70862c04d06eea9b926f4290509b8e0b09d73de7a3ea0c181811e\": container with ID starting with 64d5c9337cd70862c04d06eea9b926f4290509b8e0b09d73de7a3ea0c181811e not found: ID does not exist" containerID="64d5c9337cd70862c04d06eea9b926f4290509b8e0b09d73de7a3ea0c181811e" Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.779201 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64d5c9337cd70862c04d06eea9b926f4290509b8e0b09d73de7a3ea0c181811e"} err="failed to get container status \"64d5c9337cd70862c04d06eea9b926f4290509b8e0b09d73de7a3ea0c181811e\": rpc error: code = NotFound desc = could not find container \"64d5c9337cd70862c04d06eea9b926f4290509b8e0b09d73de7a3ea0c181811e\": container with ID starting with 64d5c9337cd70862c04d06eea9b926f4290509b8e0b09d73de7a3ea0c181811e not found: ID does not exist" Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.779235 4934 scope.go:117] "RemoveContainer" containerID="af70f69dc380a109c459550553d05155d01a8a3bbf85e90bbafd0e406fa3c047" Oct 02 10:09:37 crc kubenswrapper[4934]: E1002 10:09:37.779516 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af70f69dc380a109c459550553d05155d01a8a3bbf85e90bbafd0e406fa3c047\": container with ID starting with af70f69dc380a109c459550553d05155d01a8a3bbf85e90bbafd0e406fa3c047 not found: ID does not exist" containerID="af70f69dc380a109c459550553d05155d01a8a3bbf85e90bbafd0e406fa3c047" Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.779543 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af70f69dc380a109c459550553d05155d01a8a3bbf85e90bbafd0e406fa3c047"} err="failed to get container status \"af70f69dc380a109c459550553d05155d01a8a3bbf85e90bbafd0e406fa3c047\": rpc error: code = NotFound desc = could not find container \"af70f69dc380a109c459550553d05155d01a8a3bbf85e90bbafd0e406fa3c047\": container with ID starting with af70f69dc380a109c459550553d05155d01a8a3bbf85e90bbafd0e406fa3c047 not found: ID does not exist" Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.779562 4934 scope.go:117] "RemoveContainer" containerID="64d5c9337cd70862c04d06eea9b926f4290509b8e0b09d73de7a3ea0c181811e" Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.780201 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64d5c9337cd70862c04d06eea9b926f4290509b8e0b09d73de7a3ea0c181811e"} err="failed to get container status \"64d5c9337cd70862c04d06eea9b926f4290509b8e0b09d73de7a3ea0c181811e\": rpc error: code = NotFound desc = could not find container \"64d5c9337cd70862c04d06eea9b926f4290509b8e0b09d73de7a3ea0c181811e\": container with ID starting with 64d5c9337cd70862c04d06eea9b926f4290509b8e0b09d73de7a3ea0c181811e not found: ID does not exist" Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.780330 4934 scope.go:117] "RemoveContainer" containerID="af70f69dc380a109c459550553d05155d01a8a3bbf85e90bbafd0e406fa3c047" Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.780771 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af70f69dc380a109c459550553d05155d01a8a3bbf85e90bbafd0e406fa3c047"} err="failed to get container status \"af70f69dc380a109c459550553d05155d01a8a3bbf85e90bbafd0e406fa3c047\": rpc error: code = NotFound desc = could not find container \"af70f69dc380a109c459550553d05155d01a8a3bbf85e90bbafd0e406fa3c047\": container with ID starting with af70f69dc380a109c459550553d05155d01a8a3bbf85e90bbafd0e406fa3c047 not found: ID does not exist" Oct 02 10:09:37 crc kubenswrapper[4934]: I1002 10:09:37.802027 4934 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/17364b5a-18b8-4b54-a675-ca2c3b0d110b-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.002913 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-6c8h9" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.045400 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.051632 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.091269 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:09:38 crc kubenswrapper[4934]: E1002 10:09:38.092313 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17364b5a-18b8-4b54-a675-ca2c3b0d110b" containerName="nova-metadata-metadata" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.092333 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="17364b5a-18b8-4b54-a675-ca2c3b0d110b" containerName="nova-metadata-metadata" Oct 02 10:09:38 crc kubenswrapper[4934]: E1002 10:09:38.092350 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17364b5a-18b8-4b54-a675-ca2c3b0d110b" containerName="nova-metadata-log" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.092356 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="17364b5a-18b8-4b54-a675-ca2c3b0d110b" containerName="nova-metadata-log" Oct 02 10:09:38 crc kubenswrapper[4934]: E1002 10:09:38.092367 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed820a81-72fc-4d84-876f-9def630d4ff7" containerName="nova-manage" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.092373 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed820a81-72fc-4d84-876f-9def630d4ff7" containerName="nova-manage" Oct 02 10:09:38 crc kubenswrapper[4934]: E1002 10:09:38.092389 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73e6e086-8443-46e5-89f8-21bc9bde6b4f" containerName="init" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.092395 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="73e6e086-8443-46e5-89f8-21bc9bde6b4f" containerName="init" Oct 02 10:09:38 crc kubenswrapper[4934]: E1002 10:09:38.092418 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73e6e086-8443-46e5-89f8-21bc9bde6b4f" containerName="dnsmasq-dns" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.092424 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="73e6e086-8443-46e5-89f8-21bc9bde6b4f" containerName="dnsmasq-dns" Oct 02 10:09:38 crc kubenswrapper[4934]: E1002 10:09:38.092436 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="490173ef-a363-4b84-8802-6a69bb9bdb76" containerName="nova-cell1-conductor-db-sync" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.092442 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="490173ef-a363-4b84-8802-6a69bb9bdb76" containerName="nova-cell1-conductor-db-sync" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.092619 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed820a81-72fc-4d84-876f-9def630d4ff7" containerName="nova-manage" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.092635 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="17364b5a-18b8-4b54-a675-ca2c3b0d110b" containerName="nova-metadata-metadata" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.092649 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="73e6e086-8443-46e5-89f8-21bc9bde6b4f" containerName="dnsmasq-dns" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.092661 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="17364b5a-18b8-4b54-a675-ca2c3b0d110b" containerName="nova-metadata-log" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.092672 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="490173ef-a363-4b84-8802-6a69bb9bdb76" containerName="nova-cell1-conductor-db-sync" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.093868 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.095789 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.096737 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.107814 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/490173ef-a363-4b84-8802-6a69bb9bdb76-combined-ca-bundle\") pod \"490173ef-a363-4b84-8802-6a69bb9bdb76\" (UID: \"490173ef-a363-4b84-8802-6a69bb9bdb76\") " Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.108117 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/490173ef-a363-4b84-8802-6a69bb9bdb76-config-data\") pod \"490173ef-a363-4b84-8802-6a69bb9bdb76\" (UID: \"490173ef-a363-4b84-8802-6a69bb9bdb76\") " Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.111607 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkd92\" (UniqueName: \"kubernetes.io/projected/490173ef-a363-4b84-8802-6a69bb9bdb76-kube-api-access-qkd92\") pod \"490173ef-a363-4b84-8802-6a69bb9bdb76\" (UID: \"490173ef-a363-4b84-8802-6a69bb9bdb76\") " Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.111693 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/490173ef-a363-4b84-8802-6a69bb9bdb76-scripts\") pod \"490173ef-a363-4b84-8802-6a69bb9bdb76\" (UID: \"490173ef-a363-4b84-8802-6a69bb9bdb76\") " Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.111918 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2fpg\" (UniqueName: \"kubernetes.io/projected/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-kube-api-access-p2fpg\") pod \"nova-metadata-0\" (UID: \"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049\") " pod="openstack/nova-metadata-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.111941 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049\") " pod="openstack/nova-metadata-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.111958 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049\") " pod="openstack/nova-metadata-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.111976 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-logs\") pod \"nova-metadata-0\" (UID: \"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049\") " pod="openstack/nova-metadata-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.112071 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-config-data\") pod \"nova-metadata-0\" (UID: \"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049\") " pod="openstack/nova-metadata-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.117833 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.118753 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/490173ef-a363-4b84-8802-6a69bb9bdb76-kube-api-access-qkd92" (OuterVolumeSpecName: "kube-api-access-qkd92") pod "490173ef-a363-4b84-8802-6a69bb9bdb76" (UID: "490173ef-a363-4b84-8802-6a69bb9bdb76"). InnerVolumeSpecName "kube-api-access-qkd92". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.125330 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/490173ef-a363-4b84-8802-6a69bb9bdb76-scripts" (OuterVolumeSpecName: "scripts") pod "490173ef-a363-4b84-8802-6a69bb9bdb76" (UID: "490173ef-a363-4b84-8802-6a69bb9bdb76"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.167123 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/490173ef-a363-4b84-8802-6a69bb9bdb76-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "490173ef-a363-4b84-8802-6a69bb9bdb76" (UID: "490173ef-a363-4b84-8802-6a69bb9bdb76"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.181151 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/490173ef-a363-4b84-8802-6a69bb9bdb76-config-data" (OuterVolumeSpecName: "config-data") pod "490173ef-a363-4b84-8802-6a69bb9bdb76" (UID: "490173ef-a363-4b84-8802-6a69bb9bdb76"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.213848 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-config-data\") pod \"nova-metadata-0\" (UID: \"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049\") " pod="openstack/nova-metadata-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.213975 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2fpg\" (UniqueName: \"kubernetes.io/projected/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-kube-api-access-p2fpg\") pod \"nova-metadata-0\" (UID: \"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049\") " pod="openstack/nova-metadata-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.214007 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049\") " pod="openstack/nova-metadata-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.214031 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049\") " pod="openstack/nova-metadata-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.214058 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-logs\") pod \"nova-metadata-0\" (UID: \"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049\") " pod="openstack/nova-metadata-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.214130 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/490173ef-a363-4b84-8802-6a69bb9bdb76-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.214147 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/490173ef-a363-4b84-8802-6a69bb9bdb76-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.214162 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/490173ef-a363-4b84-8802-6a69bb9bdb76-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.214175 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkd92\" (UniqueName: \"kubernetes.io/projected/490173ef-a363-4b84-8802-6a69bb9bdb76-kube-api-access-qkd92\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.215651 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-logs\") pod \"nova-metadata-0\" (UID: \"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049\") " pod="openstack/nova-metadata-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.220082 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-config-data\") pod \"nova-metadata-0\" (UID: \"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049\") " pod="openstack/nova-metadata-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.220995 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049\") " pod="openstack/nova-metadata-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.221507 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049\") " pod="openstack/nova-metadata-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.244491 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2fpg\" (UniqueName: \"kubernetes.io/projected/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-kube-api-access-p2fpg\") pod \"nova-metadata-0\" (UID: \"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049\") " pod="openstack/nova-metadata-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.331424 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.417488 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djv4s\" (UniqueName: \"kubernetes.io/projected/06eb14a8-953f-48f9-9da1-a217d6f5568a-kube-api-access-djv4s\") pod \"06eb14a8-953f-48f9-9da1-a217d6f5568a\" (UID: \"06eb14a8-953f-48f9-9da1-a217d6f5568a\") " Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.417631 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06eb14a8-953f-48f9-9da1-a217d6f5568a-log-httpd\") pod \"06eb14a8-953f-48f9-9da1-a217d6f5568a\" (UID: \"06eb14a8-953f-48f9-9da1-a217d6f5568a\") " Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.417878 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06eb14a8-953f-48f9-9da1-a217d6f5568a-combined-ca-bundle\") pod \"06eb14a8-953f-48f9-9da1-a217d6f5568a\" (UID: \"06eb14a8-953f-48f9-9da1-a217d6f5568a\") " Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.418030 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06eb14a8-953f-48f9-9da1-a217d6f5568a-config-data\") pod \"06eb14a8-953f-48f9-9da1-a217d6f5568a\" (UID: \"06eb14a8-953f-48f9-9da1-a217d6f5568a\") " Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.418068 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06eb14a8-953f-48f9-9da1-a217d6f5568a-run-httpd\") pod \"06eb14a8-953f-48f9-9da1-a217d6f5568a\" (UID: \"06eb14a8-953f-48f9-9da1-a217d6f5568a\") " Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.418142 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/06eb14a8-953f-48f9-9da1-a217d6f5568a-sg-core-conf-yaml\") pod \"06eb14a8-953f-48f9-9da1-a217d6f5568a\" (UID: \"06eb14a8-953f-48f9-9da1-a217d6f5568a\") " Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.418178 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06eb14a8-953f-48f9-9da1-a217d6f5568a-scripts\") pod \"06eb14a8-953f-48f9-9da1-a217d6f5568a\" (UID: \"06eb14a8-953f-48f9-9da1-a217d6f5568a\") " Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.418204 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06eb14a8-953f-48f9-9da1-a217d6f5568a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "06eb14a8-953f-48f9-9da1-a217d6f5568a" (UID: "06eb14a8-953f-48f9-9da1-a217d6f5568a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.418470 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06eb14a8-953f-48f9-9da1-a217d6f5568a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "06eb14a8-953f-48f9-9da1-a217d6f5568a" (UID: "06eb14a8-953f-48f9-9da1-a217d6f5568a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.418740 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.419434 4934 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06eb14a8-953f-48f9-9da1-a217d6f5568a-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.419462 4934 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06eb14a8-953f-48f9-9da1-a217d6f5568a-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.422567 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06eb14a8-953f-48f9-9da1-a217d6f5568a-kube-api-access-djv4s" (OuterVolumeSpecName: "kube-api-access-djv4s") pod "06eb14a8-953f-48f9-9da1-a217d6f5568a" (UID: "06eb14a8-953f-48f9-9da1-a217d6f5568a"). InnerVolumeSpecName "kube-api-access-djv4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.422963 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06eb14a8-953f-48f9-9da1-a217d6f5568a-scripts" (OuterVolumeSpecName: "scripts") pod "06eb14a8-953f-48f9-9da1-a217d6f5568a" (UID: "06eb14a8-953f-48f9-9da1-a217d6f5568a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.460465 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06eb14a8-953f-48f9-9da1-a217d6f5568a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "06eb14a8-953f-48f9-9da1-a217d6f5568a" (UID: "06eb14a8-953f-48f9-9da1-a217d6f5568a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.510288 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06eb14a8-953f-48f9-9da1-a217d6f5568a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06eb14a8-953f-48f9-9da1-a217d6f5568a" (UID: "06eb14a8-953f-48f9-9da1-a217d6f5568a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.520934 4934 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/06eb14a8-953f-48f9-9da1-a217d6f5568a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.521151 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06eb14a8-953f-48f9-9da1-a217d6f5568a-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.521160 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djv4s\" (UniqueName: \"kubernetes.io/projected/06eb14a8-953f-48f9-9da1-a217d6f5568a-kube-api-access-djv4s\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.521170 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06eb14a8-953f-48f9-9da1-a217d6f5568a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.550525 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06eb14a8-953f-48f9-9da1-a217d6f5568a-config-data" (OuterVolumeSpecName: "config-data") pod "06eb14a8-953f-48f9-9da1-a217d6f5568a" (UID: "06eb14a8-953f-48f9-9da1-a217d6f5568a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.622511 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06eb14a8-953f-48f9-9da1-a217d6f5568a-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.750294 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-6c8h9" event={"ID":"490173ef-a363-4b84-8802-6a69bb9bdb76","Type":"ContainerDied","Data":"6bd88e6b84c85c775d5e2aa96af34b9bcb2b5a2ea400063eb4cf3835e68f5b28"} Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.750332 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6bd88e6b84c85c775d5e2aa96af34b9bcb2b5a2ea400063eb4cf3835e68f5b28" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.750345 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-6c8h9" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.760380 4934 generic.go:334] "Generic (PLEG): container finished" podID="06eb14a8-953f-48f9-9da1-a217d6f5568a" containerID="3a12e64befe2bba1ea2f15bb3318dd9358c70985e22da881b9157cfb0998d0a1" exitCode=0 Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.760450 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06eb14a8-953f-48f9-9da1-a217d6f5568a","Type":"ContainerDied","Data":"3a12e64befe2bba1ea2f15bb3318dd9358c70985e22da881b9157cfb0998d0a1"} Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.760481 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06eb14a8-953f-48f9-9da1-a217d6f5568a","Type":"ContainerDied","Data":"fcfcec9bc4ae0b3ad4b8ffb01c18deb1a2130bd48b2e3269ad9ada9bf468b102"} Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.760502 4934 scope.go:117] "RemoveContainer" containerID="1b6f098eddfd42439826b80bb65f1c06a66432a5934703606b4954bb0e943de3" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.760645 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.764314 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="004218d8-62bb-470f-b184-67d0ac97a27c" containerName="nova-scheduler-scheduler" containerID="cri-o://64480ebe56c42e093c07f270ecff0f024eb20a7d01a5f42839131f3edf140a1d" gracePeriod=30 Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.791274 4934 scope.go:117] "RemoveContainer" containerID="1e06b21ceb0c24b7ae0cae51c6becaf3172bbda2e4d646bf490f1cc858852347" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.804070 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 10:09:38 crc kubenswrapper[4934]: E1002 10:09:38.804477 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06eb14a8-953f-48f9-9da1-a217d6f5568a" containerName="sg-core" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.804508 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="06eb14a8-953f-48f9-9da1-a217d6f5568a" containerName="sg-core" Oct 02 10:09:38 crc kubenswrapper[4934]: E1002 10:09:38.804537 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06eb14a8-953f-48f9-9da1-a217d6f5568a" containerName="proxy-httpd" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.804545 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="06eb14a8-953f-48f9-9da1-a217d6f5568a" containerName="proxy-httpd" Oct 02 10:09:38 crc kubenswrapper[4934]: E1002 10:09:38.804568 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06eb14a8-953f-48f9-9da1-a217d6f5568a" containerName="ceilometer-notification-agent" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.804592 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="06eb14a8-953f-48f9-9da1-a217d6f5568a" containerName="ceilometer-notification-agent" Oct 02 10:09:38 crc kubenswrapper[4934]: E1002 10:09:38.804602 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06eb14a8-953f-48f9-9da1-a217d6f5568a" containerName="ceilometer-central-agent" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.804610 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="06eb14a8-953f-48f9-9da1-a217d6f5568a" containerName="ceilometer-central-agent" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.804818 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="06eb14a8-953f-48f9-9da1-a217d6f5568a" containerName="ceilometer-notification-agent" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.804840 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="06eb14a8-953f-48f9-9da1-a217d6f5568a" containerName="sg-core" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.804853 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="06eb14a8-953f-48f9-9da1-a217d6f5568a" containerName="ceilometer-central-agent" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.804870 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="06eb14a8-953f-48f9-9da1-a217d6f5568a" containerName="proxy-httpd" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.805541 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.810565 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.827588 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49a440a9-1a3f-4ee9-bf28-7996fe6f222d-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"49a440a9-1a3f-4ee9-bf28-7996fe6f222d\") " pod="openstack/nova-cell1-conductor-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.827699 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd8fm\" (UniqueName: \"kubernetes.io/projected/49a440a9-1a3f-4ee9-bf28-7996fe6f222d-kube-api-access-gd8fm\") pod \"nova-cell1-conductor-0\" (UID: \"49a440a9-1a3f-4ee9-bf28-7996fe6f222d\") " pod="openstack/nova-cell1-conductor-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.827779 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49a440a9-1a3f-4ee9-bf28-7996fe6f222d-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"49a440a9-1a3f-4ee9-bf28-7996fe6f222d\") " pod="openstack/nova-cell1-conductor-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.839021 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.844881 4934 scope.go:117] "RemoveContainer" containerID="3a12e64befe2bba1ea2f15bb3318dd9358c70985e22da881b9157cfb0998d0a1" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.861813 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.879192 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.896237 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.897764 4934 scope.go:117] "RemoveContainer" containerID="c630cb405e8c2e671298d31e62bc1dd4507dff91d4e9081ba2703b0a2c966f1a" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.899206 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.901387 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.901737 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.902629 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.931340 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f514ef27-2d57-4acd-9358-1de88b7c93cc-log-httpd\") pod \"ceilometer-0\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " pod="openstack/ceilometer-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.931503 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd8fm\" (UniqueName: \"kubernetes.io/projected/49a440a9-1a3f-4ee9-bf28-7996fe6f222d-kube-api-access-gd8fm\") pod \"nova-cell1-conductor-0\" (UID: \"49a440a9-1a3f-4ee9-bf28-7996fe6f222d\") " pod="openstack/nova-cell1-conductor-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.936025 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " pod="openstack/ceilometer-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.936058 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-scripts\") pod \"ceilometer-0\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " pod="openstack/ceilometer-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.936090 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " pod="openstack/ceilometer-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.936139 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-config-data\") pod \"ceilometer-0\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " pod="openstack/ceilometer-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.936656 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49a440a9-1a3f-4ee9-bf28-7996fe6f222d-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"49a440a9-1a3f-4ee9-bf28-7996fe6f222d\") " pod="openstack/nova-cell1-conductor-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.936783 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " pod="openstack/ceilometer-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.936824 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f514ef27-2d57-4acd-9358-1de88b7c93cc-run-httpd\") pod \"ceilometer-0\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " pod="openstack/ceilometer-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.936873 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mngf\" (UniqueName: \"kubernetes.io/projected/f514ef27-2d57-4acd-9358-1de88b7c93cc-kube-api-access-6mngf\") pod \"ceilometer-0\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " pod="openstack/ceilometer-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.936954 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49a440a9-1a3f-4ee9-bf28-7996fe6f222d-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"49a440a9-1a3f-4ee9-bf28-7996fe6f222d\") " pod="openstack/nova-cell1-conductor-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.942173 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06eb14a8-953f-48f9-9da1-a217d6f5568a" path="/var/lib/kubelet/pods/06eb14a8-953f-48f9-9da1-a217d6f5568a/volumes" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.943428 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17364b5a-18b8-4b54-a675-ca2c3b0d110b" path="/var/lib/kubelet/pods/17364b5a-18b8-4b54-a675-ca2c3b0d110b/volumes" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.944149 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.948105 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.951317 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd8fm\" (UniqueName: \"kubernetes.io/projected/49a440a9-1a3f-4ee9-bf28-7996fe6f222d-kube-api-access-gd8fm\") pod \"nova-cell1-conductor-0\" (UID: \"49a440a9-1a3f-4ee9-bf28-7996fe6f222d\") " pod="openstack/nova-cell1-conductor-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.956815 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49a440a9-1a3f-4ee9-bf28-7996fe6f222d-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"49a440a9-1a3f-4ee9-bf28-7996fe6f222d\") " pod="openstack/nova-cell1-conductor-0" Oct 02 10:09:38 crc kubenswrapper[4934]: I1002 10:09:38.958314 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49a440a9-1a3f-4ee9-bf28-7996fe6f222d-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"49a440a9-1a3f-4ee9-bf28-7996fe6f222d\") " pod="openstack/nova-cell1-conductor-0" Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.000452 4934 scope.go:117] "RemoveContainer" containerID="1b6f098eddfd42439826b80bb65f1c06a66432a5934703606b4954bb0e943de3" Oct 02 10:09:39 crc kubenswrapper[4934]: E1002 10:09:39.000909 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b6f098eddfd42439826b80bb65f1c06a66432a5934703606b4954bb0e943de3\": container with ID starting with 1b6f098eddfd42439826b80bb65f1c06a66432a5934703606b4954bb0e943de3 not found: ID does not exist" containerID="1b6f098eddfd42439826b80bb65f1c06a66432a5934703606b4954bb0e943de3" Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.000954 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b6f098eddfd42439826b80bb65f1c06a66432a5934703606b4954bb0e943de3"} err="failed to get container status \"1b6f098eddfd42439826b80bb65f1c06a66432a5934703606b4954bb0e943de3\": rpc error: code = NotFound desc = could not find container \"1b6f098eddfd42439826b80bb65f1c06a66432a5934703606b4954bb0e943de3\": container with ID starting with 1b6f098eddfd42439826b80bb65f1c06a66432a5934703606b4954bb0e943de3 not found: ID does not exist" Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.000981 4934 scope.go:117] "RemoveContainer" containerID="1e06b21ceb0c24b7ae0cae51c6becaf3172bbda2e4d646bf490f1cc858852347" Oct 02 10:09:39 crc kubenswrapper[4934]: E1002 10:09:39.001242 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e06b21ceb0c24b7ae0cae51c6becaf3172bbda2e4d646bf490f1cc858852347\": container with ID starting with 1e06b21ceb0c24b7ae0cae51c6becaf3172bbda2e4d646bf490f1cc858852347 not found: ID does not exist" containerID="1e06b21ceb0c24b7ae0cae51c6becaf3172bbda2e4d646bf490f1cc858852347" Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.001272 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e06b21ceb0c24b7ae0cae51c6becaf3172bbda2e4d646bf490f1cc858852347"} err="failed to get container status \"1e06b21ceb0c24b7ae0cae51c6becaf3172bbda2e4d646bf490f1cc858852347\": rpc error: code = NotFound desc = could not find container \"1e06b21ceb0c24b7ae0cae51c6becaf3172bbda2e4d646bf490f1cc858852347\": container with ID starting with 1e06b21ceb0c24b7ae0cae51c6becaf3172bbda2e4d646bf490f1cc858852347 not found: ID does not exist" Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.001289 4934 scope.go:117] "RemoveContainer" containerID="3a12e64befe2bba1ea2f15bb3318dd9358c70985e22da881b9157cfb0998d0a1" Oct 02 10:09:39 crc kubenswrapper[4934]: E1002 10:09:39.001469 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a12e64befe2bba1ea2f15bb3318dd9358c70985e22da881b9157cfb0998d0a1\": container with ID starting with 3a12e64befe2bba1ea2f15bb3318dd9358c70985e22da881b9157cfb0998d0a1 not found: ID does not exist" containerID="3a12e64befe2bba1ea2f15bb3318dd9358c70985e22da881b9157cfb0998d0a1" Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.001494 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a12e64befe2bba1ea2f15bb3318dd9358c70985e22da881b9157cfb0998d0a1"} err="failed to get container status \"3a12e64befe2bba1ea2f15bb3318dd9358c70985e22da881b9157cfb0998d0a1\": rpc error: code = NotFound desc = could not find container \"3a12e64befe2bba1ea2f15bb3318dd9358c70985e22da881b9157cfb0998d0a1\": container with ID starting with 3a12e64befe2bba1ea2f15bb3318dd9358c70985e22da881b9157cfb0998d0a1 not found: ID does not exist" Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.001511 4934 scope.go:117] "RemoveContainer" containerID="c630cb405e8c2e671298d31e62bc1dd4507dff91d4e9081ba2703b0a2c966f1a" Oct 02 10:09:39 crc kubenswrapper[4934]: E1002 10:09:39.001780 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c630cb405e8c2e671298d31e62bc1dd4507dff91d4e9081ba2703b0a2c966f1a\": container with ID starting with c630cb405e8c2e671298d31e62bc1dd4507dff91d4e9081ba2703b0a2c966f1a not found: ID does not exist" containerID="c630cb405e8c2e671298d31e62bc1dd4507dff91d4e9081ba2703b0a2c966f1a" Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.001808 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c630cb405e8c2e671298d31e62bc1dd4507dff91d4e9081ba2703b0a2c966f1a"} err="failed to get container status \"c630cb405e8c2e671298d31e62bc1dd4507dff91d4e9081ba2703b0a2c966f1a\": rpc error: code = NotFound desc = could not find container \"c630cb405e8c2e671298d31e62bc1dd4507dff91d4e9081ba2703b0a2c966f1a\": container with ID starting with c630cb405e8c2e671298d31e62bc1dd4507dff91d4e9081ba2703b0a2c966f1a not found: ID does not exist" Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.040351 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " pod="openstack/ceilometer-0" Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.040415 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f514ef27-2d57-4acd-9358-1de88b7c93cc-run-httpd\") pod \"ceilometer-0\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " pod="openstack/ceilometer-0" Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.040448 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mngf\" (UniqueName: \"kubernetes.io/projected/f514ef27-2d57-4acd-9358-1de88b7c93cc-kube-api-access-6mngf\") pod \"ceilometer-0\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " pod="openstack/ceilometer-0" Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.040540 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f514ef27-2d57-4acd-9358-1de88b7c93cc-log-httpd\") pod \"ceilometer-0\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " pod="openstack/ceilometer-0" Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.040624 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " pod="openstack/ceilometer-0" Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.040646 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-scripts\") pod \"ceilometer-0\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " pod="openstack/ceilometer-0" Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.040681 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " pod="openstack/ceilometer-0" Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.040708 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-config-data\") pod \"ceilometer-0\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " pod="openstack/ceilometer-0" Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.041453 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f514ef27-2d57-4acd-9358-1de88b7c93cc-log-httpd\") pod \"ceilometer-0\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " pod="openstack/ceilometer-0" Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.041624 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f514ef27-2d57-4acd-9358-1de88b7c93cc-run-httpd\") pod \"ceilometer-0\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " pod="openstack/ceilometer-0" Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.044835 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " pod="openstack/ceilometer-0" Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.044939 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-scripts\") pod \"ceilometer-0\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " pod="openstack/ceilometer-0" Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.046960 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " pod="openstack/ceilometer-0" Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.047564 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-config-data\") pod \"ceilometer-0\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " pod="openstack/ceilometer-0" Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.060308 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " pod="openstack/ceilometer-0" Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.062897 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mngf\" (UniqueName: \"kubernetes.io/projected/f514ef27-2d57-4acd-9358-1de88b7c93cc-kube-api-access-6mngf\") pod \"ceilometer-0\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " pod="openstack/ceilometer-0" Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.140111 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.299097 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.569052 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:09:39 crc kubenswrapper[4934]: W1002 10:09:39.583943 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf514ef27_2d57_4acd_9358_1de88b7c93cc.slice/crio-8b7205a7e9ac9eb1bd97b55ce961a2bff25332584ac8717b8244f13b5d759cd0 WatchSource:0}: Error finding container 8b7205a7e9ac9eb1bd97b55ce961a2bff25332584ac8717b8244f13b5d759cd0: Status 404 returned error can't find the container with id 8b7205a7e9ac9eb1bd97b55ce961a2bff25332584ac8717b8244f13b5d759cd0 Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.628614 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.780009 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f514ef27-2d57-4acd-9358-1de88b7c93cc","Type":"ContainerStarted","Data":"8b7205a7e9ac9eb1bd97b55ce961a2bff25332584ac8717b8244f13b5d759cd0"} Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.783748 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"49a440a9-1a3f-4ee9-bf28-7996fe6f222d","Type":"ContainerStarted","Data":"0ae4d557135649d0f5b152d89b3a85b31e941838defa90173e4ff571d93acb6e"} Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.786119 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049","Type":"ContainerStarted","Data":"2d1cfcb32101f27ad01e6c51bb7828a926d69c41b6d7fa00b917776ca6b15cc4"} Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.786165 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049","Type":"ContainerStarted","Data":"de95cb4af5f7758289a9ed99c925c143cf2617fd14807a19a8624ee861cf0c14"} Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.786177 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049","Type":"ContainerStarted","Data":"fcf0e04f0d8939aa100d8c0efadcf5f98a2780516f6570c44a389946090c1ea6"} Oct 02 10:09:39 crc kubenswrapper[4934]: I1002 10:09:39.811854 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.811832557 podStartE2EDuration="1.811832557s" podCreationTimestamp="2025-10-02 10:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:09:39.806184151 +0000 UTC m=+1251.558825673" watchObservedRunningTime="2025-10-02 10:09:39.811832557 +0000 UTC m=+1251.564474079" Oct 02 10:09:40 crc kubenswrapper[4934]: E1002 10:09:40.177846 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="64480ebe56c42e093c07f270ecff0f024eb20a7d01a5f42839131f3edf140a1d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 10:09:40 crc kubenswrapper[4934]: E1002 10:09:40.179839 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="64480ebe56c42e093c07f270ecff0f024eb20a7d01a5f42839131f3edf140a1d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 10:09:40 crc kubenswrapper[4934]: E1002 10:09:40.181389 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="64480ebe56c42e093c07f270ecff0f024eb20a7d01a5f42839131f3edf140a1d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 10:09:40 crc kubenswrapper[4934]: E1002 10:09:40.181427 4934 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="004218d8-62bb-470f-b184-67d0ac97a27c" containerName="nova-scheduler-scheduler" Oct 02 10:09:40 crc kubenswrapper[4934]: I1002 10:09:40.813311 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f514ef27-2d57-4acd-9358-1de88b7c93cc","Type":"ContainerStarted","Data":"7b0e6add9855a99cdda02553879be12b877a2d36f3263bddfc03546b6fcf64a4"} Oct 02 10:09:40 crc kubenswrapper[4934]: I1002 10:09:40.815679 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"49a440a9-1a3f-4ee9-bf28-7996fe6f222d","Type":"ContainerStarted","Data":"38289b907213ccb98a8b74c9827dd591276a04d2139306b7c6ab58968debd430"} Oct 02 10:09:40 crc kubenswrapper[4934]: I1002 10:09:40.832462 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.8324431519999997 podStartE2EDuration="2.832443152s" podCreationTimestamp="2025-10-02 10:09:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:09:40.830279422 +0000 UTC m=+1252.582920954" watchObservedRunningTime="2025-10-02 10:09:40.832443152 +0000 UTC m=+1252.585084674" Oct 02 10:09:41 crc kubenswrapper[4934]: I1002 10:09:41.828809 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f514ef27-2d57-4acd-9358-1de88b7c93cc","Type":"ContainerStarted","Data":"377c8d141638babfca9fabaf0a76592a05ac9a05623a93af4743eb00f4b4387e"} Oct 02 10:09:41 crc kubenswrapper[4934]: I1002 10:09:41.829082 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f514ef27-2d57-4acd-9358-1de88b7c93cc","Type":"ContainerStarted","Data":"71d0fe1f529c97ac51b9d6f3dff9dcd2e0a6ea93cd5d7fc9c5a39464cf0b0ded"} Oct 02 10:09:41 crc kubenswrapper[4934]: I1002 10:09:41.829101 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 02 10:09:42 crc kubenswrapper[4934]: I1002 10:09:42.826942 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:09:42 crc kubenswrapper[4934]: I1002 10:09:42.839391 4934 generic.go:334] "Generic (PLEG): container finished" podID="004218d8-62bb-470f-b184-67d0ac97a27c" containerID="64480ebe56c42e093c07f270ecff0f024eb20a7d01a5f42839131f3edf140a1d" exitCode=0 Oct 02 10:09:42 crc kubenswrapper[4934]: I1002 10:09:42.839470 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"004218d8-62bb-470f-b184-67d0ac97a27c","Type":"ContainerDied","Data":"64480ebe56c42e093c07f270ecff0f024eb20a7d01a5f42839131f3edf140a1d"} Oct 02 10:09:42 crc kubenswrapper[4934]: I1002 10:09:42.845666 4934 generic.go:334] "Generic (PLEG): container finished" podID="b4da04f7-1f59-4acc-90e0-7b7cba140393" containerID="a6d1113b52ffb7001a8b2d59c298de5526ba2211698784ccbcdf6bb49e0e2e01" exitCode=0 Oct 02 10:09:42 crc kubenswrapper[4934]: I1002 10:09:42.846648 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b4da04f7-1f59-4acc-90e0-7b7cba140393","Type":"ContainerDied","Data":"a6d1113b52ffb7001a8b2d59c298de5526ba2211698784ccbcdf6bb49e0e2e01"} Oct 02 10:09:42 crc kubenswrapper[4934]: I1002 10:09:42.846709 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b4da04f7-1f59-4acc-90e0-7b7cba140393","Type":"ContainerDied","Data":"ddc17ca02bee1b3765b3561124fc12110839ea616bc262bfe97657e44dbf103c"} Oct 02 10:09:42 crc kubenswrapper[4934]: I1002 10:09:42.846730 4934 scope.go:117] "RemoveContainer" containerID="a6d1113b52ffb7001a8b2d59c298de5526ba2211698784ccbcdf6bb49e0e2e01" Oct 02 10:09:42 crc kubenswrapper[4934]: I1002 10:09:42.848561 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:09:42 crc kubenswrapper[4934]: I1002 10:09:42.900060 4934 scope.go:117] "RemoveContainer" containerID="2b6fe7b50c803ac8ece2433dd06ec23dcfef9b8a9b399c435d0bd26a0fc0e9f2" Oct 02 10:09:42 crc kubenswrapper[4934]: I1002 10:09:42.915272 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4da04f7-1f59-4acc-90e0-7b7cba140393-config-data\") pod \"b4da04f7-1f59-4acc-90e0-7b7cba140393\" (UID: \"b4da04f7-1f59-4acc-90e0-7b7cba140393\") " Oct 02 10:09:42 crc kubenswrapper[4934]: I1002 10:09:42.915323 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4da04f7-1f59-4acc-90e0-7b7cba140393-combined-ca-bundle\") pod \"b4da04f7-1f59-4acc-90e0-7b7cba140393\" (UID: \"b4da04f7-1f59-4acc-90e0-7b7cba140393\") " Oct 02 10:09:42 crc kubenswrapper[4934]: I1002 10:09:42.915416 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlb44\" (UniqueName: \"kubernetes.io/projected/b4da04f7-1f59-4acc-90e0-7b7cba140393-kube-api-access-vlb44\") pod \"b4da04f7-1f59-4acc-90e0-7b7cba140393\" (UID: \"b4da04f7-1f59-4acc-90e0-7b7cba140393\") " Oct 02 10:09:42 crc kubenswrapper[4934]: I1002 10:09:42.915442 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4da04f7-1f59-4acc-90e0-7b7cba140393-logs\") pod \"b4da04f7-1f59-4acc-90e0-7b7cba140393\" (UID: \"b4da04f7-1f59-4acc-90e0-7b7cba140393\") " Oct 02 10:09:42 crc kubenswrapper[4934]: I1002 10:09:42.921280 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4da04f7-1f59-4acc-90e0-7b7cba140393-logs" (OuterVolumeSpecName: "logs") pod "b4da04f7-1f59-4acc-90e0-7b7cba140393" (UID: "b4da04f7-1f59-4acc-90e0-7b7cba140393"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:09:42 crc kubenswrapper[4934]: I1002 10:09:42.948312 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4da04f7-1f59-4acc-90e0-7b7cba140393-kube-api-access-vlb44" (OuterVolumeSpecName: "kube-api-access-vlb44") pod "b4da04f7-1f59-4acc-90e0-7b7cba140393" (UID: "b4da04f7-1f59-4acc-90e0-7b7cba140393"). InnerVolumeSpecName "kube-api-access-vlb44". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:09:42 crc kubenswrapper[4934]: I1002 10:09:42.965663 4934 scope.go:117] "RemoveContainer" containerID="a6d1113b52ffb7001a8b2d59c298de5526ba2211698784ccbcdf6bb49e0e2e01" Oct 02 10:09:42 crc kubenswrapper[4934]: E1002 10:09:42.966519 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6d1113b52ffb7001a8b2d59c298de5526ba2211698784ccbcdf6bb49e0e2e01\": container with ID starting with a6d1113b52ffb7001a8b2d59c298de5526ba2211698784ccbcdf6bb49e0e2e01 not found: ID does not exist" containerID="a6d1113b52ffb7001a8b2d59c298de5526ba2211698784ccbcdf6bb49e0e2e01" Oct 02 10:09:42 crc kubenswrapper[4934]: I1002 10:09:42.966559 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6d1113b52ffb7001a8b2d59c298de5526ba2211698784ccbcdf6bb49e0e2e01"} err="failed to get container status \"a6d1113b52ffb7001a8b2d59c298de5526ba2211698784ccbcdf6bb49e0e2e01\": rpc error: code = NotFound desc = could not find container \"a6d1113b52ffb7001a8b2d59c298de5526ba2211698784ccbcdf6bb49e0e2e01\": container with ID starting with a6d1113b52ffb7001a8b2d59c298de5526ba2211698784ccbcdf6bb49e0e2e01 not found: ID does not exist" Oct 02 10:09:42 crc kubenswrapper[4934]: I1002 10:09:42.966614 4934 scope.go:117] "RemoveContainer" containerID="2b6fe7b50c803ac8ece2433dd06ec23dcfef9b8a9b399c435d0bd26a0fc0e9f2" Oct 02 10:09:42 crc kubenswrapper[4934]: E1002 10:09:42.967899 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b6fe7b50c803ac8ece2433dd06ec23dcfef9b8a9b399c435d0bd26a0fc0e9f2\": container with ID starting with 2b6fe7b50c803ac8ece2433dd06ec23dcfef9b8a9b399c435d0bd26a0fc0e9f2 not found: ID does not exist" containerID="2b6fe7b50c803ac8ece2433dd06ec23dcfef9b8a9b399c435d0bd26a0fc0e9f2" Oct 02 10:09:42 crc kubenswrapper[4934]: I1002 10:09:42.968009 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b6fe7b50c803ac8ece2433dd06ec23dcfef9b8a9b399c435d0bd26a0fc0e9f2"} err="failed to get container status \"2b6fe7b50c803ac8ece2433dd06ec23dcfef9b8a9b399c435d0bd26a0fc0e9f2\": rpc error: code = NotFound desc = could not find container \"2b6fe7b50c803ac8ece2433dd06ec23dcfef9b8a9b399c435d0bd26a0fc0e9f2\": container with ID starting with 2b6fe7b50c803ac8ece2433dd06ec23dcfef9b8a9b399c435d0bd26a0fc0e9f2 not found: ID does not exist" Oct 02 10:09:42 crc kubenswrapper[4934]: I1002 10:09:42.984925 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4da04f7-1f59-4acc-90e0-7b7cba140393-config-data" (OuterVolumeSpecName: "config-data") pod "b4da04f7-1f59-4acc-90e0-7b7cba140393" (UID: "b4da04f7-1f59-4acc-90e0-7b7cba140393"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.001861 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4da04f7-1f59-4acc-90e0-7b7cba140393-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b4da04f7-1f59-4acc-90e0-7b7cba140393" (UID: "b4da04f7-1f59-4acc-90e0-7b7cba140393"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.020280 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlb44\" (UniqueName: \"kubernetes.io/projected/b4da04f7-1f59-4acc-90e0-7b7cba140393-kube-api-access-vlb44\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.020314 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4da04f7-1f59-4acc-90e0-7b7cba140393-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.020328 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4da04f7-1f59-4acc-90e0-7b7cba140393-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.020340 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4da04f7-1f59-4acc-90e0-7b7cba140393-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.045218 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.053684 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.121242 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/004218d8-62bb-470f-b184-67d0ac97a27c-config-data\") pod \"004218d8-62bb-470f-b184-67d0ac97a27c\" (UID: \"004218d8-62bb-470f-b184-67d0ac97a27c\") " Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.121292 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/004218d8-62bb-470f-b184-67d0ac97a27c-combined-ca-bundle\") pod \"004218d8-62bb-470f-b184-67d0ac97a27c\" (UID: \"004218d8-62bb-470f-b184-67d0ac97a27c\") " Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.121428 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkdm2\" (UniqueName: \"kubernetes.io/projected/004218d8-62bb-470f-b184-67d0ac97a27c-kube-api-access-pkdm2\") pod \"004218d8-62bb-470f-b184-67d0ac97a27c\" (UID: \"004218d8-62bb-470f-b184-67d0ac97a27c\") " Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.127228 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/004218d8-62bb-470f-b184-67d0ac97a27c-kube-api-access-pkdm2" (OuterVolumeSpecName: "kube-api-access-pkdm2") pod "004218d8-62bb-470f-b184-67d0ac97a27c" (UID: "004218d8-62bb-470f-b184-67d0ac97a27c"). InnerVolumeSpecName "kube-api-access-pkdm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.152626 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/004218d8-62bb-470f-b184-67d0ac97a27c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "004218d8-62bb-470f-b184-67d0ac97a27c" (UID: "004218d8-62bb-470f-b184-67d0ac97a27c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.179541 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/004218d8-62bb-470f-b184-67d0ac97a27c-config-data" (OuterVolumeSpecName: "config-data") pod "004218d8-62bb-470f-b184-67d0ac97a27c" (UID: "004218d8-62bb-470f-b184-67d0ac97a27c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.199021 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.221108 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.223230 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/004218d8-62bb-470f-b184-67d0ac97a27c-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.223255 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/004218d8-62bb-470f-b184-67d0ac97a27c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.223266 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkdm2\" (UniqueName: \"kubernetes.io/projected/004218d8-62bb-470f-b184-67d0ac97a27c-kube-api-access-pkdm2\") on node \"crc\" DevicePath \"\"" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.243893 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 10:09:43 crc kubenswrapper[4934]: E1002 10:09:43.244322 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="004218d8-62bb-470f-b184-67d0ac97a27c" containerName="nova-scheduler-scheduler" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.244340 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="004218d8-62bb-470f-b184-67d0ac97a27c" containerName="nova-scheduler-scheduler" Oct 02 10:09:43 crc kubenswrapper[4934]: E1002 10:09:43.244355 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4da04f7-1f59-4acc-90e0-7b7cba140393" containerName="nova-api-log" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.244361 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4da04f7-1f59-4acc-90e0-7b7cba140393" containerName="nova-api-log" Oct 02 10:09:43 crc kubenswrapper[4934]: E1002 10:09:43.244383 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4da04f7-1f59-4acc-90e0-7b7cba140393" containerName="nova-api-api" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.244389 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4da04f7-1f59-4acc-90e0-7b7cba140393" containerName="nova-api-api" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.244551 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="004218d8-62bb-470f-b184-67d0ac97a27c" containerName="nova-scheduler-scheduler" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.244565 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4da04f7-1f59-4acc-90e0-7b7cba140393" containerName="nova-api-api" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.244656 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4da04f7-1f59-4acc-90e0-7b7cba140393" containerName="nova-api-log" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.245826 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.248348 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.249605 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.324329 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-799fg\" (UniqueName: \"kubernetes.io/projected/6475956e-9f6b-4af0-a1fd-32dc920aae5f-kube-api-access-799fg\") pod \"nova-api-0\" (UID: \"6475956e-9f6b-4af0-a1fd-32dc920aae5f\") " pod="openstack/nova-api-0" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.324392 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6475956e-9f6b-4af0-a1fd-32dc920aae5f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6475956e-9f6b-4af0-a1fd-32dc920aae5f\") " pod="openstack/nova-api-0" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.324451 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6475956e-9f6b-4af0-a1fd-32dc920aae5f-logs\") pod \"nova-api-0\" (UID: \"6475956e-9f6b-4af0-a1fd-32dc920aae5f\") " pod="openstack/nova-api-0" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.324523 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6475956e-9f6b-4af0-a1fd-32dc920aae5f-config-data\") pod \"nova-api-0\" (UID: \"6475956e-9f6b-4af0-a1fd-32dc920aae5f\") " pod="openstack/nova-api-0" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.420226 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.420302 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.426150 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-799fg\" (UniqueName: \"kubernetes.io/projected/6475956e-9f6b-4af0-a1fd-32dc920aae5f-kube-api-access-799fg\") pod \"nova-api-0\" (UID: \"6475956e-9f6b-4af0-a1fd-32dc920aae5f\") " pod="openstack/nova-api-0" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.426221 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6475956e-9f6b-4af0-a1fd-32dc920aae5f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6475956e-9f6b-4af0-a1fd-32dc920aae5f\") " pod="openstack/nova-api-0" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.426293 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6475956e-9f6b-4af0-a1fd-32dc920aae5f-logs\") pod \"nova-api-0\" (UID: \"6475956e-9f6b-4af0-a1fd-32dc920aae5f\") " pod="openstack/nova-api-0" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.426348 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6475956e-9f6b-4af0-a1fd-32dc920aae5f-config-data\") pod \"nova-api-0\" (UID: \"6475956e-9f6b-4af0-a1fd-32dc920aae5f\") " pod="openstack/nova-api-0" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.426916 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6475956e-9f6b-4af0-a1fd-32dc920aae5f-logs\") pod \"nova-api-0\" (UID: \"6475956e-9f6b-4af0-a1fd-32dc920aae5f\") " pod="openstack/nova-api-0" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.430934 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6475956e-9f6b-4af0-a1fd-32dc920aae5f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6475956e-9f6b-4af0-a1fd-32dc920aae5f\") " pod="openstack/nova-api-0" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.440097 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6475956e-9f6b-4af0-a1fd-32dc920aae5f-config-data\") pod \"nova-api-0\" (UID: \"6475956e-9f6b-4af0-a1fd-32dc920aae5f\") " pod="openstack/nova-api-0" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.446878 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-799fg\" (UniqueName: \"kubernetes.io/projected/6475956e-9f6b-4af0-a1fd-32dc920aae5f-kube-api-access-799fg\") pod \"nova-api-0\" (UID: \"6475956e-9f6b-4af0-a1fd-32dc920aae5f\") " pod="openstack/nova-api-0" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.578995 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.854315 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.854314 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"004218d8-62bb-470f-b184-67d0ac97a27c","Type":"ContainerDied","Data":"1632b8483ea033cbbba2c7c067242ff1fd2f18a267115927a0e9d79df8b1f449"} Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.854721 4934 scope.go:117] "RemoveContainer" containerID="64480ebe56c42e093c07f270ecff0f024eb20a7d01a5f42839131f3edf140a1d" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.857092 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f514ef27-2d57-4acd-9358-1de88b7c93cc","Type":"ContainerStarted","Data":"d4eeba7c708cd3a06d41746960438bc56385f7816339ff060ce5f18fae6e6b40"} Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.857241 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.880134 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.35891077 podStartE2EDuration="5.880114066s" podCreationTimestamp="2025-10-02 10:09:38 +0000 UTC" firstStartedPulling="2025-10-02 10:09:39.586309111 +0000 UTC m=+1251.338950633" lastFinishedPulling="2025-10-02 10:09:43.107512407 +0000 UTC m=+1254.860153929" observedRunningTime="2025-10-02 10:09:43.874184732 +0000 UTC m=+1255.626826254" watchObservedRunningTime="2025-10-02 10:09:43.880114066 +0000 UTC m=+1255.632755588" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.898880 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.915421 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.926123 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.928415 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.933062 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 10:09:43 crc kubenswrapper[4934]: I1002 10:09:43.937745 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:09:44 crc kubenswrapper[4934]: I1002 10:09:44.038476 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eeb3879e-b8fc-43ab-ad26-42ae80715706-config-data\") pod \"nova-scheduler-0\" (UID: \"eeb3879e-b8fc-43ab-ad26-42ae80715706\") " pod="openstack/nova-scheduler-0" Oct 02 10:09:44 crc kubenswrapper[4934]: I1002 10:09:44.038924 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb3879e-b8fc-43ab-ad26-42ae80715706-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"eeb3879e-b8fc-43ab-ad26-42ae80715706\") " pod="openstack/nova-scheduler-0" Oct 02 10:09:44 crc kubenswrapper[4934]: I1002 10:09:44.038951 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-th89f\" (UniqueName: \"kubernetes.io/projected/eeb3879e-b8fc-43ab-ad26-42ae80715706-kube-api-access-th89f\") pod \"nova-scheduler-0\" (UID: \"eeb3879e-b8fc-43ab-ad26-42ae80715706\") " pod="openstack/nova-scheduler-0" Oct 02 10:09:44 crc kubenswrapper[4934]: I1002 10:09:44.046126 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:09:44 crc kubenswrapper[4934]: I1002 10:09:44.140459 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb3879e-b8fc-43ab-ad26-42ae80715706-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"eeb3879e-b8fc-43ab-ad26-42ae80715706\") " pod="openstack/nova-scheduler-0" Oct 02 10:09:44 crc kubenswrapper[4934]: I1002 10:09:44.140504 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-th89f\" (UniqueName: \"kubernetes.io/projected/eeb3879e-b8fc-43ab-ad26-42ae80715706-kube-api-access-th89f\") pod \"nova-scheduler-0\" (UID: \"eeb3879e-b8fc-43ab-ad26-42ae80715706\") " pod="openstack/nova-scheduler-0" Oct 02 10:09:44 crc kubenswrapper[4934]: I1002 10:09:44.140593 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eeb3879e-b8fc-43ab-ad26-42ae80715706-config-data\") pod \"nova-scheduler-0\" (UID: \"eeb3879e-b8fc-43ab-ad26-42ae80715706\") " pod="openstack/nova-scheduler-0" Oct 02 10:09:44 crc kubenswrapper[4934]: I1002 10:09:44.150285 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb3879e-b8fc-43ab-ad26-42ae80715706-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"eeb3879e-b8fc-43ab-ad26-42ae80715706\") " pod="openstack/nova-scheduler-0" Oct 02 10:09:44 crc kubenswrapper[4934]: I1002 10:09:44.158065 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eeb3879e-b8fc-43ab-ad26-42ae80715706-config-data\") pod \"nova-scheduler-0\" (UID: \"eeb3879e-b8fc-43ab-ad26-42ae80715706\") " pod="openstack/nova-scheduler-0" Oct 02 10:09:44 crc kubenswrapper[4934]: I1002 10:09:44.173154 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-th89f\" (UniqueName: \"kubernetes.io/projected/eeb3879e-b8fc-43ab-ad26-42ae80715706-kube-api-access-th89f\") pod \"nova-scheduler-0\" (UID: \"eeb3879e-b8fc-43ab-ad26-42ae80715706\") " pod="openstack/nova-scheduler-0" Oct 02 10:09:44 crc kubenswrapper[4934]: I1002 10:09:44.246153 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 02 10:09:44 crc kubenswrapper[4934]: I1002 10:09:44.260707 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 10:09:44 crc kubenswrapper[4934]: I1002 10:09:44.700716 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:09:44 crc kubenswrapper[4934]: I1002 10:09:44.874990 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6475956e-9f6b-4af0-a1fd-32dc920aae5f","Type":"ContainerStarted","Data":"87a457ba083a6014c881b6560d691cee748ee0b4576186204ec9cedbada2b5bd"} Oct 02 10:09:44 crc kubenswrapper[4934]: I1002 10:09:44.875039 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6475956e-9f6b-4af0-a1fd-32dc920aae5f","Type":"ContainerStarted","Data":"dcd299d59c44300718908764238c92a2012fcf7a687961e5cb069658954c374c"} Oct 02 10:09:44 crc kubenswrapper[4934]: I1002 10:09:44.875056 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6475956e-9f6b-4af0-a1fd-32dc920aae5f","Type":"ContainerStarted","Data":"914edb394ba3dc3c9a4b7c07f69ca067becc59058c06c0368b811fd89bea60b2"} Oct 02 10:09:44 crc kubenswrapper[4934]: I1002 10:09:44.880950 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"eeb3879e-b8fc-43ab-ad26-42ae80715706","Type":"ContainerStarted","Data":"a14bbcf611a2af141e70f921257f49fbe1eb653bd690ef1c6acab5a466d9415d"} Oct 02 10:09:44 crc kubenswrapper[4934]: I1002 10:09:44.896264 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.896245037 podStartE2EDuration="1.896245037s" podCreationTimestamp="2025-10-02 10:09:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:09:44.895683601 +0000 UTC m=+1256.648325133" watchObservedRunningTime="2025-10-02 10:09:44.896245037 +0000 UTC m=+1256.648886559" Oct 02 10:09:44 crc kubenswrapper[4934]: I1002 10:09:44.926946 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="004218d8-62bb-470f-b184-67d0ac97a27c" path="/var/lib/kubelet/pods/004218d8-62bb-470f-b184-67d0ac97a27c/volumes" Oct 02 10:09:44 crc kubenswrapper[4934]: I1002 10:09:44.927643 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4da04f7-1f59-4acc-90e0-7b7cba140393" path="/var/lib/kubelet/pods/b4da04f7-1f59-4acc-90e0-7b7cba140393/volumes" Oct 02 10:09:45 crc kubenswrapper[4934]: I1002 10:09:45.893825 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"eeb3879e-b8fc-43ab-ad26-42ae80715706","Type":"ContainerStarted","Data":"c6b414c90f3b2c8ee4f300f9213c49d4f93007c4ebe815709d38d734fde7cf3b"} Oct 02 10:09:45 crc kubenswrapper[4934]: I1002 10:09:45.916403 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.916382719 podStartE2EDuration="2.916382719s" podCreationTimestamp="2025-10-02 10:09:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:09:45.913798788 +0000 UTC m=+1257.666440330" watchObservedRunningTime="2025-10-02 10:09:45.916382719 +0000 UTC m=+1257.669024241" Oct 02 10:09:48 crc kubenswrapper[4934]: I1002 10:09:48.420631 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 10:09:48 crc kubenswrapper[4934]: I1002 10:09:48.420974 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 10:09:49 crc kubenswrapper[4934]: I1002 10:09:49.261631 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 10:09:49 crc kubenswrapper[4934]: I1002 10:09:49.431756 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="edb4b618-5f25-4a83-a2e7-3b2e2d6a2049" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.192:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 10:09:49 crc kubenswrapper[4934]: I1002 10:09:49.432161 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="edb4b618-5f25-4a83-a2e7-3b2e2d6a2049" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.192:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 10:09:53 crc kubenswrapper[4934]: I1002 10:09:53.579343 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 10:09:53 crc kubenswrapper[4934]: I1002 10:09:53.580141 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 10:09:54 crc kubenswrapper[4934]: I1002 10:09:54.261496 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 10:09:54 crc kubenswrapper[4934]: I1002 10:09:54.293719 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 10:09:54 crc kubenswrapper[4934]: I1002 10:09:54.661860 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6475956e-9f6b-4af0-a1fd-32dc920aae5f" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.195:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 10:09:54 crc kubenswrapper[4934]: I1002 10:09:54.661863 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6475956e-9f6b-4af0-a1fd-32dc920aae5f" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.195:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 10:09:55 crc kubenswrapper[4934]: I1002 10:09:55.048964 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 10:09:58 crc kubenswrapper[4934]: I1002 10:09:58.427473 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 10:09:58 crc kubenswrapper[4934]: I1002 10:09:58.430981 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 10:09:58 crc kubenswrapper[4934]: I1002 10:09:58.435502 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 10:09:59 crc kubenswrapper[4934]: I1002 10:09:59.073959 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.027040 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.091333 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e92017-966a-499d-8bb5-3225aacd6dd7-combined-ca-bundle\") pod \"21e92017-966a-499d-8bb5-3225aacd6dd7\" (UID: \"21e92017-966a-499d-8bb5-3225aacd6dd7\") " Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.091528 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e92017-966a-499d-8bb5-3225aacd6dd7-config-data\") pod \"21e92017-966a-499d-8bb5-3225aacd6dd7\" (UID: \"21e92017-966a-499d-8bb5-3225aacd6dd7\") " Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.091791 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8fqf\" (UniqueName: \"kubernetes.io/projected/21e92017-966a-499d-8bb5-3225aacd6dd7-kube-api-access-l8fqf\") pod \"21e92017-966a-499d-8bb5-3225aacd6dd7\" (UID: \"21e92017-966a-499d-8bb5-3225aacd6dd7\") " Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.097877 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21e92017-966a-499d-8bb5-3225aacd6dd7-kube-api-access-l8fqf" (OuterVolumeSpecName: "kube-api-access-l8fqf") pod "21e92017-966a-499d-8bb5-3225aacd6dd7" (UID: "21e92017-966a-499d-8bb5-3225aacd6dd7"). InnerVolumeSpecName "kube-api-access-l8fqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.100964 4934 generic.go:334] "Generic (PLEG): container finished" podID="21e92017-966a-499d-8bb5-3225aacd6dd7" containerID="494857a62fbea1b42d73d034dc76d537a8a6f515d0153bc68199f0984a52453d" exitCode=137 Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.101035 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"21e92017-966a-499d-8bb5-3225aacd6dd7","Type":"ContainerDied","Data":"494857a62fbea1b42d73d034dc76d537a8a6f515d0153bc68199f0984a52453d"} Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.101066 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"21e92017-966a-499d-8bb5-3225aacd6dd7","Type":"ContainerDied","Data":"7766f03265fb0fd733e4d2139d3d9e86739a072eb857d065a3c448ee5bb64cdc"} Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.101111 4934 scope.go:117] "RemoveContainer" containerID="494857a62fbea1b42d73d034dc76d537a8a6f515d0153bc68199f0984a52453d" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.101302 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.129150 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21e92017-966a-499d-8bb5-3225aacd6dd7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "21e92017-966a-499d-8bb5-3225aacd6dd7" (UID: "21e92017-966a-499d-8bb5-3225aacd6dd7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.141815 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21e92017-966a-499d-8bb5-3225aacd6dd7-config-data" (OuterVolumeSpecName: "config-data") pod "21e92017-966a-499d-8bb5-3225aacd6dd7" (UID: "21e92017-966a-499d-8bb5-3225aacd6dd7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.193876 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8fqf\" (UniqueName: \"kubernetes.io/projected/21e92017-966a-499d-8bb5-3225aacd6dd7-kube-api-access-l8fqf\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.193921 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e92017-966a-499d-8bb5-3225aacd6dd7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.193933 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e92017-966a-499d-8bb5-3225aacd6dd7-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.209946 4934 scope.go:117] "RemoveContainer" containerID="494857a62fbea1b42d73d034dc76d537a8a6f515d0153bc68199f0984a52453d" Oct 02 10:10:02 crc kubenswrapper[4934]: E1002 10:10:02.210394 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"494857a62fbea1b42d73d034dc76d537a8a6f515d0153bc68199f0984a52453d\": container with ID starting with 494857a62fbea1b42d73d034dc76d537a8a6f515d0153bc68199f0984a52453d not found: ID does not exist" containerID="494857a62fbea1b42d73d034dc76d537a8a6f515d0153bc68199f0984a52453d" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.210461 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"494857a62fbea1b42d73d034dc76d537a8a6f515d0153bc68199f0984a52453d"} err="failed to get container status \"494857a62fbea1b42d73d034dc76d537a8a6f515d0153bc68199f0984a52453d\": rpc error: code = NotFound desc = could not find container \"494857a62fbea1b42d73d034dc76d537a8a6f515d0153bc68199f0984a52453d\": container with ID starting with 494857a62fbea1b42d73d034dc76d537a8a6f515d0153bc68199f0984a52453d not found: ID does not exist" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.437695 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.450256 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.463698 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 10:10:02 crc kubenswrapper[4934]: E1002 10:10:02.464112 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21e92017-966a-499d-8bb5-3225aacd6dd7" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.464138 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="21e92017-966a-499d-8bb5-3225aacd6dd7" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.464382 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="21e92017-966a-499d-8bb5-3225aacd6dd7" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.465168 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.468233 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.468673 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.481736 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.489356 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.601807 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.601881 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.602180 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzczs\" (UniqueName: \"kubernetes.io/projected/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-kube-api-access-mzczs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.602275 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.602426 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.704024 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzczs\" (UniqueName: \"kubernetes.io/projected/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-kube-api-access-mzczs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.704083 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.704118 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.704173 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.704197 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.712076 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.713398 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.713507 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.715713 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.733658 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzczs\" (UniqueName: \"kubernetes.io/projected/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-kube-api-access-mzczs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.790068 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:10:02 crc kubenswrapper[4934]: I1002 10:10:02.950423 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21e92017-966a-499d-8bb5-3225aacd6dd7" path="/var/lib/kubelet/pods/21e92017-966a-499d-8bb5-3225aacd6dd7/volumes" Oct 02 10:10:03 crc kubenswrapper[4934]: I1002 10:10:03.300958 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 10:10:03 crc kubenswrapper[4934]: W1002 10:10:03.308086 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f37c71e_afcc_44c5_aa1b_bf8fb6d8af8f.slice/crio-0697d241dfeac3d026f3b9bbabc143e393f12b6a58f2588141d9ee2fd2b2b1f0 WatchSource:0}: Error finding container 0697d241dfeac3d026f3b9bbabc143e393f12b6a58f2588141d9ee2fd2b2b1f0: Status 404 returned error can't find the container with id 0697d241dfeac3d026f3b9bbabc143e393f12b6a58f2588141d9ee2fd2b2b1f0 Oct 02 10:10:03 crc kubenswrapper[4934]: I1002 10:10:03.585390 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 10:10:03 crc kubenswrapper[4934]: I1002 10:10:03.586339 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 10:10:03 crc kubenswrapper[4934]: I1002 10:10:03.588686 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 10:10:03 crc kubenswrapper[4934]: I1002 10:10:03.591187 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 10:10:04 crc kubenswrapper[4934]: I1002 10:10:04.125007 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f","Type":"ContainerStarted","Data":"b9baba9779ef141c5b9d9cb5a3a64ef9893f0e264bbcdb5320d4acc16116a004"} Oct 02 10:10:04 crc kubenswrapper[4934]: I1002 10:10:04.125086 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f","Type":"ContainerStarted","Data":"0697d241dfeac3d026f3b9bbabc143e393f12b6a58f2588141d9ee2fd2b2b1f0"} Oct 02 10:10:04 crc kubenswrapper[4934]: I1002 10:10:04.125334 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 10:10:04 crc kubenswrapper[4934]: I1002 10:10:04.135025 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 10:10:04 crc kubenswrapper[4934]: I1002 10:10:04.158069 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.158047508 podStartE2EDuration="2.158047508s" podCreationTimestamp="2025-10-02 10:10:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:10:04.146686494 +0000 UTC m=+1275.899328036" watchObservedRunningTime="2025-10-02 10:10:04.158047508 +0000 UTC m=+1275.910689050" Oct 02 10:10:04 crc kubenswrapper[4934]: I1002 10:10:04.347915 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85b4cf56ff-zr6h2"] Oct 02 10:10:04 crc kubenswrapper[4934]: I1002 10:10:04.349418 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" Oct 02 10:10:04 crc kubenswrapper[4934]: I1002 10:10:04.397448 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85b4cf56ff-zr6h2"] Oct 02 10:10:04 crc kubenswrapper[4934]: I1002 10:10:04.443147 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-ovsdbserver-sb\") pod \"dnsmasq-dns-85b4cf56ff-zr6h2\" (UID: \"20b39d07-1559-4d11-b9ff-6ef97142a58a\") " pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" Oct 02 10:10:04 crc kubenswrapper[4934]: I1002 10:10:04.443233 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-config\") pod \"dnsmasq-dns-85b4cf56ff-zr6h2\" (UID: \"20b39d07-1559-4d11-b9ff-6ef97142a58a\") " pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" Oct 02 10:10:04 crc kubenswrapper[4934]: I1002 10:10:04.443301 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-dns-swift-storage-0\") pod \"dnsmasq-dns-85b4cf56ff-zr6h2\" (UID: \"20b39d07-1559-4d11-b9ff-6ef97142a58a\") " pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" Oct 02 10:10:04 crc kubenswrapper[4934]: I1002 10:10:04.443387 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65775\" (UniqueName: \"kubernetes.io/projected/20b39d07-1559-4d11-b9ff-6ef97142a58a-kube-api-access-65775\") pod \"dnsmasq-dns-85b4cf56ff-zr6h2\" (UID: \"20b39d07-1559-4d11-b9ff-6ef97142a58a\") " pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" Oct 02 10:10:04 crc kubenswrapper[4934]: I1002 10:10:04.443432 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-dns-svc\") pod \"dnsmasq-dns-85b4cf56ff-zr6h2\" (UID: \"20b39d07-1559-4d11-b9ff-6ef97142a58a\") " pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" Oct 02 10:10:04 crc kubenswrapper[4934]: I1002 10:10:04.443477 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-ovsdbserver-nb\") pod \"dnsmasq-dns-85b4cf56ff-zr6h2\" (UID: \"20b39d07-1559-4d11-b9ff-6ef97142a58a\") " pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" Oct 02 10:10:04 crc kubenswrapper[4934]: I1002 10:10:04.544751 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-ovsdbserver-sb\") pod \"dnsmasq-dns-85b4cf56ff-zr6h2\" (UID: \"20b39d07-1559-4d11-b9ff-6ef97142a58a\") " pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" Oct 02 10:10:04 crc kubenswrapper[4934]: I1002 10:10:04.544842 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-config\") pod \"dnsmasq-dns-85b4cf56ff-zr6h2\" (UID: \"20b39d07-1559-4d11-b9ff-6ef97142a58a\") " pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" Oct 02 10:10:04 crc kubenswrapper[4934]: I1002 10:10:04.544861 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-dns-swift-storage-0\") pod \"dnsmasq-dns-85b4cf56ff-zr6h2\" (UID: \"20b39d07-1559-4d11-b9ff-6ef97142a58a\") " pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" Oct 02 10:10:04 crc kubenswrapper[4934]: I1002 10:10:04.544923 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65775\" (UniqueName: \"kubernetes.io/projected/20b39d07-1559-4d11-b9ff-6ef97142a58a-kube-api-access-65775\") pod \"dnsmasq-dns-85b4cf56ff-zr6h2\" (UID: \"20b39d07-1559-4d11-b9ff-6ef97142a58a\") " pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" Oct 02 10:10:04 crc kubenswrapper[4934]: I1002 10:10:04.544939 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-dns-svc\") pod \"dnsmasq-dns-85b4cf56ff-zr6h2\" (UID: \"20b39d07-1559-4d11-b9ff-6ef97142a58a\") " pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" Oct 02 10:10:04 crc kubenswrapper[4934]: I1002 10:10:04.544985 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-ovsdbserver-nb\") pod \"dnsmasq-dns-85b4cf56ff-zr6h2\" (UID: \"20b39d07-1559-4d11-b9ff-6ef97142a58a\") " pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" Oct 02 10:10:04 crc kubenswrapper[4934]: I1002 10:10:04.545683 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-ovsdbserver-sb\") pod \"dnsmasq-dns-85b4cf56ff-zr6h2\" (UID: \"20b39d07-1559-4d11-b9ff-6ef97142a58a\") " pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" Oct 02 10:10:04 crc kubenswrapper[4934]: I1002 10:10:04.545923 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-dns-swift-storage-0\") pod \"dnsmasq-dns-85b4cf56ff-zr6h2\" (UID: \"20b39d07-1559-4d11-b9ff-6ef97142a58a\") " pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" Oct 02 10:10:04 crc kubenswrapper[4934]: I1002 10:10:04.545961 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-dns-svc\") pod \"dnsmasq-dns-85b4cf56ff-zr6h2\" (UID: \"20b39d07-1559-4d11-b9ff-6ef97142a58a\") " pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" Oct 02 10:10:04 crc kubenswrapper[4934]: I1002 10:10:04.546459 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-config\") pod \"dnsmasq-dns-85b4cf56ff-zr6h2\" (UID: \"20b39d07-1559-4d11-b9ff-6ef97142a58a\") " pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" Oct 02 10:10:04 crc kubenswrapper[4934]: I1002 10:10:04.547289 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-ovsdbserver-nb\") pod \"dnsmasq-dns-85b4cf56ff-zr6h2\" (UID: \"20b39d07-1559-4d11-b9ff-6ef97142a58a\") " pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" Oct 02 10:10:04 crc kubenswrapper[4934]: I1002 10:10:04.574161 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65775\" (UniqueName: \"kubernetes.io/projected/20b39d07-1559-4d11-b9ff-6ef97142a58a-kube-api-access-65775\") pod \"dnsmasq-dns-85b4cf56ff-zr6h2\" (UID: \"20b39d07-1559-4d11-b9ff-6ef97142a58a\") " pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" Oct 02 10:10:04 crc kubenswrapper[4934]: I1002 10:10:04.692561 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" Oct 02 10:10:05 crc kubenswrapper[4934]: I1002 10:10:05.180812 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85b4cf56ff-zr6h2"] Oct 02 10:10:05 crc kubenswrapper[4934]: W1002 10:10:05.183595 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20b39d07_1559_4d11_b9ff_6ef97142a58a.slice/crio-a2f8d8a5389fc04137cb0d450fa6ab57ff03fe6a761dfa3a3bad6b4d03aa3c82 WatchSource:0}: Error finding container a2f8d8a5389fc04137cb0d450fa6ab57ff03fe6a761dfa3a3bad6b4d03aa3c82: Status 404 returned error can't find the container with id a2f8d8a5389fc04137cb0d450fa6ab57ff03fe6a761dfa3a3bad6b4d03aa3c82 Oct 02 10:10:06 crc kubenswrapper[4934]: I1002 10:10:06.142653 4934 generic.go:334] "Generic (PLEG): container finished" podID="20b39d07-1559-4d11-b9ff-6ef97142a58a" containerID="6ae422d564bec96f2ecabdd62c9fb7ad073e005254728994f68b5086ec8bc468" exitCode=0 Oct 02 10:10:06 crc kubenswrapper[4934]: I1002 10:10:06.142884 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" event={"ID":"20b39d07-1559-4d11-b9ff-6ef97142a58a","Type":"ContainerDied","Data":"6ae422d564bec96f2ecabdd62c9fb7ad073e005254728994f68b5086ec8bc468"} Oct 02 10:10:06 crc kubenswrapper[4934]: I1002 10:10:06.143340 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" event={"ID":"20b39d07-1559-4d11-b9ff-6ef97142a58a","Type":"ContainerStarted","Data":"a2f8d8a5389fc04137cb0d450fa6ab57ff03fe6a761dfa3a3bad6b4d03aa3c82"} Oct 02 10:10:06 crc kubenswrapper[4934]: I1002 10:10:06.379698 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:10:06 crc kubenswrapper[4934]: I1002 10:10:06.380012 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f514ef27-2d57-4acd-9358-1de88b7c93cc" containerName="ceilometer-central-agent" containerID="cri-o://7b0e6add9855a99cdda02553879be12b877a2d36f3263bddfc03546b6fcf64a4" gracePeriod=30 Oct 02 10:10:06 crc kubenswrapper[4934]: I1002 10:10:06.380143 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f514ef27-2d57-4acd-9358-1de88b7c93cc" containerName="proxy-httpd" containerID="cri-o://d4eeba7c708cd3a06d41746960438bc56385f7816339ff060ce5f18fae6e6b40" gracePeriod=30 Oct 02 10:10:06 crc kubenswrapper[4934]: I1002 10:10:06.380188 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f514ef27-2d57-4acd-9358-1de88b7c93cc" containerName="sg-core" containerID="cri-o://377c8d141638babfca9fabaf0a76592a05ac9a05623a93af4743eb00f4b4387e" gracePeriod=30 Oct 02 10:10:06 crc kubenswrapper[4934]: I1002 10:10:06.380218 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f514ef27-2d57-4acd-9358-1de88b7c93cc" containerName="ceilometer-notification-agent" containerID="cri-o://71d0fe1f529c97ac51b9d6f3dff9dcd2e0a6ea93cd5d7fc9c5a39464cf0b0ded" gracePeriod=30 Oct 02 10:10:06 crc kubenswrapper[4934]: I1002 10:10:06.392231 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="f514ef27-2d57-4acd-9358-1de88b7c93cc" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.158688 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" event={"ID":"20b39d07-1559-4d11-b9ff-6ef97142a58a","Type":"ContainerStarted","Data":"aac633fecdf45004f95036aeac08948cfcfe75a04ef1942aef60308cc2356517"} Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.159068 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.167203 4934 generic.go:334] "Generic (PLEG): container finished" podID="f514ef27-2d57-4acd-9358-1de88b7c93cc" containerID="d4eeba7c708cd3a06d41746960438bc56385f7816339ff060ce5f18fae6e6b40" exitCode=0 Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.167236 4934 generic.go:334] "Generic (PLEG): container finished" podID="f514ef27-2d57-4acd-9358-1de88b7c93cc" containerID="377c8d141638babfca9fabaf0a76592a05ac9a05623a93af4743eb00f4b4387e" exitCode=2 Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.167247 4934 generic.go:334] "Generic (PLEG): container finished" podID="f514ef27-2d57-4acd-9358-1de88b7c93cc" containerID="71d0fe1f529c97ac51b9d6f3dff9dcd2e0a6ea93cd5d7fc9c5a39464cf0b0ded" exitCode=0 Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.167257 4934 generic.go:334] "Generic (PLEG): container finished" podID="f514ef27-2d57-4acd-9358-1de88b7c93cc" containerID="7b0e6add9855a99cdda02553879be12b877a2d36f3263bddfc03546b6fcf64a4" exitCode=0 Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.167283 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f514ef27-2d57-4acd-9358-1de88b7c93cc","Type":"ContainerDied","Data":"d4eeba7c708cd3a06d41746960438bc56385f7816339ff060ce5f18fae6e6b40"} Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.167312 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f514ef27-2d57-4acd-9358-1de88b7c93cc","Type":"ContainerDied","Data":"377c8d141638babfca9fabaf0a76592a05ac9a05623a93af4743eb00f4b4387e"} Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.167324 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f514ef27-2d57-4acd-9358-1de88b7c93cc","Type":"ContainerDied","Data":"71d0fe1f529c97ac51b9d6f3dff9dcd2e0a6ea93cd5d7fc9c5a39464cf0b0ded"} Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.167337 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f514ef27-2d57-4acd-9358-1de88b7c93cc","Type":"ContainerDied","Data":"7b0e6add9855a99cdda02553879be12b877a2d36f3263bddfc03546b6fcf64a4"} Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.190382 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" podStartSLOduration=3.190360968 podStartE2EDuration="3.190360968s" podCreationTimestamp="2025-10-02 10:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:10:07.179359035 +0000 UTC m=+1278.932000557" watchObservedRunningTime="2025-10-02 10:10:07.190360968 +0000 UTC m=+1278.943002490" Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.280934 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.281483 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6475956e-9f6b-4af0-a1fd-32dc920aae5f" containerName="nova-api-log" containerID="cri-o://dcd299d59c44300718908764238c92a2012fcf7a687961e5cb069658954c374c" gracePeriod=30 Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.281655 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6475956e-9f6b-4af0-a1fd-32dc920aae5f" containerName="nova-api-api" containerID="cri-o://87a457ba083a6014c881b6560d691cee748ee0b4576186204ec9cedbada2b5bd" gracePeriod=30 Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.471480 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.602559 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-config-data\") pod \"f514ef27-2d57-4acd-9358-1de88b7c93cc\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.602617 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-sg-core-conf-yaml\") pod \"f514ef27-2d57-4acd-9358-1de88b7c93cc\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.602685 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f514ef27-2d57-4acd-9358-1de88b7c93cc-log-httpd\") pod \"f514ef27-2d57-4acd-9358-1de88b7c93cc\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.602758 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mngf\" (UniqueName: \"kubernetes.io/projected/f514ef27-2d57-4acd-9358-1de88b7c93cc-kube-api-access-6mngf\") pod \"f514ef27-2d57-4acd-9358-1de88b7c93cc\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.602788 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-scripts\") pod \"f514ef27-2d57-4acd-9358-1de88b7c93cc\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.602914 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-combined-ca-bundle\") pod \"f514ef27-2d57-4acd-9358-1de88b7c93cc\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.602941 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-ceilometer-tls-certs\") pod \"f514ef27-2d57-4acd-9358-1de88b7c93cc\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.602987 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f514ef27-2d57-4acd-9358-1de88b7c93cc-run-httpd\") pod \"f514ef27-2d57-4acd-9358-1de88b7c93cc\" (UID: \"f514ef27-2d57-4acd-9358-1de88b7c93cc\") " Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.603271 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f514ef27-2d57-4acd-9358-1de88b7c93cc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f514ef27-2d57-4acd-9358-1de88b7c93cc" (UID: "f514ef27-2d57-4acd-9358-1de88b7c93cc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.603486 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f514ef27-2d57-4acd-9358-1de88b7c93cc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f514ef27-2d57-4acd-9358-1de88b7c93cc" (UID: "f514ef27-2d57-4acd-9358-1de88b7c93cc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.603617 4934 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f514ef27-2d57-4acd-9358-1de88b7c93cc-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.603631 4934 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f514ef27-2d57-4acd-9358-1de88b7c93cc-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.612295 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-scripts" (OuterVolumeSpecName: "scripts") pod "f514ef27-2d57-4acd-9358-1de88b7c93cc" (UID: "f514ef27-2d57-4acd-9358-1de88b7c93cc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.612494 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f514ef27-2d57-4acd-9358-1de88b7c93cc-kube-api-access-6mngf" (OuterVolumeSpecName: "kube-api-access-6mngf") pod "f514ef27-2d57-4acd-9358-1de88b7c93cc" (UID: "f514ef27-2d57-4acd-9358-1de88b7c93cc"). InnerVolumeSpecName "kube-api-access-6mngf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.642826 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f514ef27-2d57-4acd-9358-1de88b7c93cc" (UID: "f514ef27-2d57-4acd-9358-1de88b7c93cc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.668488 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "f514ef27-2d57-4acd-9358-1de88b7c93cc" (UID: "f514ef27-2d57-4acd-9358-1de88b7c93cc"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.706668 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mngf\" (UniqueName: \"kubernetes.io/projected/f514ef27-2d57-4acd-9358-1de88b7c93cc-kube-api-access-6mngf\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.706714 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.706729 4934 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.706740 4934 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.710429 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f514ef27-2d57-4acd-9358-1de88b7c93cc" (UID: "f514ef27-2d57-4acd-9358-1de88b7c93cc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.737685 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-config-data" (OuterVolumeSpecName: "config-data") pod "f514ef27-2d57-4acd-9358-1de88b7c93cc" (UID: "f514ef27-2d57-4acd-9358-1de88b7c93cc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.790412 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.808244 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:07 crc kubenswrapper[4934]: I1002 10:10:07.808284 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f514ef27-2d57-4acd-9358-1de88b7c93cc-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.179091 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f514ef27-2d57-4acd-9358-1de88b7c93cc","Type":"ContainerDied","Data":"8b7205a7e9ac9eb1bd97b55ce961a2bff25332584ac8717b8244f13b5d759cd0"} Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.179137 4934 scope.go:117] "RemoveContainer" containerID="d4eeba7c708cd3a06d41746960438bc56385f7816339ff060ce5f18fae6e6b40" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.179142 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.181135 4934 generic.go:334] "Generic (PLEG): container finished" podID="6475956e-9f6b-4af0-a1fd-32dc920aae5f" containerID="dcd299d59c44300718908764238c92a2012fcf7a687961e5cb069658954c374c" exitCode=143 Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.181946 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6475956e-9f6b-4af0-a1fd-32dc920aae5f","Type":"ContainerDied","Data":"dcd299d59c44300718908764238c92a2012fcf7a687961e5cb069658954c374c"} Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.220006 4934 scope.go:117] "RemoveContainer" containerID="377c8d141638babfca9fabaf0a76592a05ac9a05623a93af4743eb00f4b4387e" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.221063 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.246681 4934 scope.go:117] "RemoveContainer" containerID="71d0fe1f529c97ac51b9d6f3dff9dcd2e0a6ea93cd5d7fc9c5a39464cf0b0ded" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.276675 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.289434 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.290962 4934 scope.go:117] "RemoveContainer" containerID="7b0e6add9855a99cdda02553879be12b877a2d36f3263bddfc03546b6fcf64a4" Oct 02 10:10:08 crc kubenswrapper[4934]: E1002 10:10:08.304360 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f514ef27-2d57-4acd-9358-1de88b7c93cc" containerName="sg-core" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.304414 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f514ef27-2d57-4acd-9358-1de88b7c93cc" containerName="sg-core" Oct 02 10:10:08 crc kubenswrapper[4934]: E1002 10:10:08.304443 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f514ef27-2d57-4acd-9358-1de88b7c93cc" containerName="ceilometer-central-agent" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.304451 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f514ef27-2d57-4acd-9358-1de88b7c93cc" containerName="ceilometer-central-agent" Oct 02 10:10:08 crc kubenswrapper[4934]: E1002 10:10:08.304471 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f514ef27-2d57-4acd-9358-1de88b7c93cc" containerName="proxy-httpd" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.304476 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f514ef27-2d57-4acd-9358-1de88b7c93cc" containerName="proxy-httpd" Oct 02 10:10:08 crc kubenswrapper[4934]: E1002 10:10:08.304490 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f514ef27-2d57-4acd-9358-1de88b7c93cc" containerName="ceilometer-notification-agent" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.304496 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f514ef27-2d57-4acd-9358-1de88b7c93cc" containerName="ceilometer-notification-agent" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.304722 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f514ef27-2d57-4acd-9358-1de88b7c93cc" containerName="sg-core" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.304735 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f514ef27-2d57-4acd-9358-1de88b7c93cc" containerName="ceilometer-notification-agent" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.304757 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f514ef27-2d57-4acd-9358-1de88b7c93cc" containerName="proxy-httpd" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.304768 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f514ef27-2d57-4acd-9358-1de88b7c93cc" containerName="ceilometer-central-agent" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.306344 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.306432 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.309089 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.309152 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.309629 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.429088 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-config-data\") pod \"ceilometer-0\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " pod="openstack/ceilometer-0" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.429133 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6937abdf-169a-4562-8bb4-e54204c7ce14-run-httpd\") pod \"ceilometer-0\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " pod="openstack/ceilometer-0" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.429158 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " pod="openstack/ceilometer-0" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.429184 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-scripts\") pod \"ceilometer-0\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " pod="openstack/ceilometer-0" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.429231 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " pod="openstack/ceilometer-0" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.429264 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " pod="openstack/ceilometer-0" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.429312 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6937abdf-169a-4562-8bb4-e54204c7ce14-log-httpd\") pod \"ceilometer-0\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " pod="openstack/ceilometer-0" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.429386 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mgm8\" (UniqueName: \"kubernetes.io/projected/6937abdf-169a-4562-8bb4-e54204c7ce14-kube-api-access-2mgm8\") pod \"ceilometer-0\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " pod="openstack/ceilometer-0" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.439643 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.439689 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.531234 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mgm8\" (UniqueName: \"kubernetes.io/projected/6937abdf-169a-4562-8bb4-e54204c7ce14-kube-api-access-2mgm8\") pod \"ceilometer-0\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " pod="openstack/ceilometer-0" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.531346 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-config-data\") pod \"ceilometer-0\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " pod="openstack/ceilometer-0" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.531370 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6937abdf-169a-4562-8bb4-e54204c7ce14-run-httpd\") pod \"ceilometer-0\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " pod="openstack/ceilometer-0" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.531394 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " pod="openstack/ceilometer-0" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.531420 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-scripts\") pod \"ceilometer-0\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " pod="openstack/ceilometer-0" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.531469 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " pod="openstack/ceilometer-0" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.531502 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " pod="openstack/ceilometer-0" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.531555 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6937abdf-169a-4562-8bb4-e54204c7ce14-log-httpd\") pod \"ceilometer-0\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " pod="openstack/ceilometer-0" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.532178 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6937abdf-169a-4562-8bb4-e54204c7ce14-log-httpd\") pod \"ceilometer-0\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " pod="openstack/ceilometer-0" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.533039 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6937abdf-169a-4562-8bb4-e54204c7ce14-run-httpd\") pod \"ceilometer-0\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " pod="openstack/ceilometer-0" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.538379 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " pod="openstack/ceilometer-0" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.538455 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-scripts\") pod \"ceilometer-0\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " pod="openstack/ceilometer-0" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.538806 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-config-data\") pod \"ceilometer-0\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " pod="openstack/ceilometer-0" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.542615 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " pod="openstack/ceilometer-0" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.545129 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " pod="openstack/ceilometer-0" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.555721 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mgm8\" (UniqueName: \"kubernetes.io/projected/6937abdf-169a-4562-8bb4-e54204c7ce14-kube-api-access-2mgm8\") pod \"ceilometer-0\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " pod="openstack/ceilometer-0" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.627787 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:10:08 crc kubenswrapper[4934]: I1002 10:10:08.933542 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f514ef27-2d57-4acd-9358-1de88b7c93cc" path="/var/lib/kubelet/pods/f514ef27-2d57-4acd-9358-1de88b7c93cc/volumes" Oct 02 10:10:09 crc kubenswrapper[4934]: I1002 10:10:09.078027 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:10:09 crc kubenswrapper[4934]: I1002 10:10:09.190915 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6937abdf-169a-4562-8bb4-e54204c7ce14","Type":"ContainerStarted","Data":"9b5db17d56e63b66861f15357b776757240748e143fef0ce789beacc510d0d9b"} Oct 02 10:10:09 crc kubenswrapper[4934]: I1002 10:10:09.852644 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:10:10 crc kubenswrapper[4934]: I1002 10:10:10.204755 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6937abdf-169a-4562-8bb4-e54204c7ce14","Type":"ContainerStarted","Data":"8af720ba645d7f4ba050ad0a52b9c6e4c81b6e54491755c67365693a836d27a2"} Oct 02 10:10:10 crc kubenswrapper[4934]: I1002 10:10:10.911999 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:10:10 crc kubenswrapper[4934]: I1002 10:10:10.985655 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6475956e-9f6b-4af0-a1fd-32dc920aae5f-logs\") pod \"6475956e-9f6b-4af0-a1fd-32dc920aae5f\" (UID: \"6475956e-9f6b-4af0-a1fd-32dc920aae5f\") " Oct 02 10:10:10 crc kubenswrapper[4934]: I1002 10:10:10.985726 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6475956e-9f6b-4af0-a1fd-32dc920aae5f-combined-ca-bundle\") pod \"6475956e-9f6b-4af0-a1fd-32dc920aae5f\" (UID: \"6475956e-9f6b-4af0-a1fd-32dc920aae5f\") " Oct 02 10:10:10 crc kubenswrapper[4934]: I1002 10:10:10.985808 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6475956e-9f6b-4af0-a1fd-32dc920aae5f-config-data\") pod \"6475956e-9f6b-4af0-a1fd-32dc920aae5f\" (UID: \"6475956e-9f6b-4af0-a1fd-32dc920aae5f\") " Oct 02 10:10:10 crc kubenswrapper[4934]: I1002 10:10:10.985868 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-799fg\" (UniqueName: \"kubernetes.io/projected/6475956e-9f6b-4af0-a1fd-32dc920aae5f-kube-api-access-799fg\") pod \"6475956e-9f6b-4af0-a1fd-32dc920aae5f\" (UID: \"6475956e-9f6b-4af0-a1fd-32dc920aae5f\") " Oct 02 10:10:10 crc kubenswrapper[4934]: I1002 10:10:10.988649 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6475956e-9f6b-4af0-a1fd-32dc920aae5f-logs" (OuterVolumeSpecName: "logs") pod "6475956e-9f6b-4af0-a1fd-32dc920aae5f" (UID: "6475956e-9f6b-4af0-a1fd-32dc920aae5f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:10:10 crc kubenswrapper[4934]: I1002 10:10:10.993080 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6475956e-9f6b-4af0-a1fd-32dc920aae5f-kube-api-access-799fg" (OuterVolumeSpecName: "kube-api-access-799fg") pod "6475956e-9f6b-4af0-a1fd-32dc920aae5f" (UID: "6475956e-9f6b-4af0-a1fd-32dc920aae5f"). InnerVolumeSpecName "kube-api-access-799fg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.021986 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6475956e-9f6b-4af0-a1fd-32dc920aae5f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6475956e-9f6b-4af0-a1fd-32dc920aae5f" (UID: "6475956e-9f6b-4af0-a1fd-32dc920aae5f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.027722 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6475956e-9f6b-4af0-a1fd-32dc920aae5f-config-data" (OuterVolumeSpecName: "config-data") pod "6475956e-9f6b-4af0-a1fd-32dc920aae5f" (UID: "6475956e-9f6b-4af0-a1fd-32dc920aae5f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.088796 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6475956e-9f6b-4af0-a1fd-32dc920aae5f-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.088850 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6475956e-9f6b-4af0-a1fd-32dc920aae5f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.088862 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6475956e-9f6b-4af0-a1fd-32dc920aae5f-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.088872 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-799fg\" (UniqueName: \"kubernetes.io/projected/6475956e-9f6b-4af0-a1fd-32dc920aae5f-kube-api-access-799fg\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.219765 4934 generic.go:334] "Generic (PLEG): container finished" podID="6475956e-9f6b-4af0-a1fd-32dc920aae5f" containerID="87a457ba083a6014c881b6560d691cee748ee0b4576186204ec9cedbada2b5bd" exitCode=0 Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.219842 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6475956e-9f6b-4af0-a1fd-32dc920aae5f","Type":"ContainerDied","Data":"87a457ba083a6014c881b6560d691cee748ee0b4576186204ec9cedbada2b5bd"} Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.219868 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6475956e-9f6b-4af0-a1fd-32dc920aae5f","Type":"ContainerDied","Data":"914edb394ba3dc3c9a4b7c07f69ca067becc59058c06c0368b811fd89bea60b2"} Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.219884 4934 scope.go:117] "RemoveContainer" containerID="87a457ba083a6014c881b6560d691cee748ee0b4576186204ec9cedbada2b5bd" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.220020 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.228898 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6937abdf-169a-4562-8bb4-e54204c7ce14","Type":"ContainerStarted","Data":"da90aa2914849d19265600ca300f389c7251f44f36c42ef0c3d0178ade6962d2"} Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.246833 4934 scope.go:117] "RemoveContainer" containerID="dcd299d59c44300718908764238c92a2012fcf7a687961e5cb069658954c374c" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.263737 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.266528 4934 scope.go:117] "RemoveContainer" containerID="87a457ba083a6014c881b6560d691cee748ee0b4576186204ec9cedbada2b5bd" Oct 02 10:10:11 crc kubenswrapper[4934]: E1002 10:10:11.267128 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87a457ba083a6014c881b6560d691cee748ee0b4576186204ec9cedbada2b5bd\": container with ID starting with 87a457ba083a6014c881b6560d691cee748ee0b4576186204ec9cedbada2b5bd not found: ID does not exist" containerID="87a457ba083a6014c881b6560d691cee748ee0b4576186204ec9cedbada2b5bd" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.267175 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87a457ba083a6014c881b6560d691cee748ee0b4576186204ec9cedbada2b5bd"} err="failed to get container status \"87a457ba083a6014c881b6560d691cee748ee0b4576186204ec9cedbada2b5bd\": rpc error: code = NotFound desc = could not find container \"87a457ba083a6014c881b6560d691cee748ee0b4576186204ec9cedbada2b5bd\": container with ID starting with 87a457ba083a6014c881b6560d691cee748ee0b4576186204ec9cedbada2b5bd not found: ID does not exist" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.267210 4934 scope.go:117] "RemoveContainer" containerID="dcd299d59c44300718908764238c92a2012fcf7a687961e5cb069658954c374c" Oct 02 10:10:11 crc kubenswrapper[4934]: E1002 10:10:11.267607 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcd299d59c44300718908764238c92a2012fcf7a687961e5cb069658954c374c\": container with ID starting with dcd299d59c44300718908764238c92a2012fcf7a687961e5cb069658954c374c not found: ID does not exist" containerID="dcd299d59c44300718908764238c92a2012fcf7a687961e5cb069658954c374c" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.267631 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcd299d59c44300718908764238c92a2012fcf7a687961e5cb069658954c374c"} err="failed to get container status \"dcd299d59c44300718908764238c92a2012fcf7a687961e5cb069658954c374c\": rpc error: code = NotFound desc = could not find container \"dcd299d59c44300718908764238c92a2012fcf7a687961e5cb069658954c374c\": container with ID starting with dcd299d59c44300718908764238c92a2012fcf7a687961e5cb069658954c374c not found: ID does not exist" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.272457 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.284285 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 10:10:11 crc kubenswrapper[4934]: E1002 10:10:11.284676 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6475956e-9f6b-4af0-a1fd-32dc920aae5f" containerName="nova-api-api" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.284694 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="6475956e-9f6b-4af0-a1fd-32dc920aae5f" containerName="nova-api-api" Oct 02 10:10:11 crc kubenswrapper[4934]: E1002 10:10:11.284715 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6475956e-9f6b-4af0-a1fd-32dc920aae5f" containerName="nova-api-log" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.284721 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="6475956e-9f6b-4af0-a1fd-32dc920aae5f" containerName="nova-api-log" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.285542 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="6475956e-9f6b-4af0-a1fd-32dc920aae5f" containerName="nova-api-log" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.285585 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="6475956e-9f6b-4af0-a1fd-32dc920aae5f" containerName="nova-api-api" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.286482 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.288367 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.288437 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.289065 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.317030 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.393586 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/585be2c9-64da-4716-9992-00e1c28380e5-config-data\") pod \"nova-api-0\" (UID: \"585be2c9-64da-4716-9992-00e1c28380e5\") " pod="openstack/nova-api-0" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.393872 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585be2c9-64da-4716-9992-00e1c28380e5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"585be2c9-64da-4716-9992-00e1c28380e5\") " pod="openstack/nova-api-0" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.393901 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/585be2c9-64da-4716-9992-00e1c28380e5-logs\") pod \"nova-api-0\" (UID: \"585be2c9-64da-4716-9992-00e1c28380e5\") " pod="openstack/nova-api-0" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.393947 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/585be2c9-64da-4716-9992-00e1c28380e5-public-tls-certs\") pod \"nova-api-0\" (UID: \"585be2c9-64da-4716-9992-00e1c28380e5\") " pod="openstack/nova-api-0" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.394017 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58f8m\" (UniqueName: \"kubernetes.io/projected/585be2c9-64da-4716-9992-00e1c28380e5-kube-api-access-58f8m\") pod \"nova-api-0\" (UID: \"585be2c9-64da-4716-9992-00e1c28380e5\") " pod="openstack/nova-api-0" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.394035 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/585be2c9-64da-4716-9992-00e1c28380e5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"585be2c9-64da-4716-9992-00e1c28380e5\") " pod="openstack/nova-api-0" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.496125 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58f8m\" (UniqueName: \"kubernetes.io/projected/585be2c9-64da-4716-9992-00e1c28380e5-kube-api-access-58f8m\") pod \"nova-api-0\" (UID: \"585be2c9-64da-4716-9992-00e1c28380e5\") " pod="openstack/nova-api-0" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.496184 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/585be2c9-64da-4716-9992-00e1c28380e5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"585be2c9-64da-4716-9992-00e1c28380e5\") " pod="openstack/nova-api-0" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.496347 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/585be2c9-64da-4716-9992-00e1c28380e5-config-data\") pod \"nova-api-0\" (UID: \"585be2c9-64da-4716-9992-00e1c28380e5\") " pod="openstack/nova-api-0" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.496418 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585be2c9-64da-4716-9992-00e1c28380e5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"585be2c9-64da-4716-9992-00e1c28380e5\") " pod="openstack/nova-api-0" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.496464 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/585be2c9-64da-4716-9992-00e1c28380e5-logs\") pod \"nova-api-0\" (UID: \"585be2c9-64da-4716-9992-00e1c28380e5\") " pod="openstack/nova-api-0" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.496556 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/585be2c9-64da-4716-9992-00e1c28380e5-public-tls-certs\") pod \"nova-api-0\" (UID: \"585be2c9-64da-4716-9992-00e1c28380e5\") " pod="openstack/nova-api-0" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.497086 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/585be2c9-64da-4716-9992-00e1c28380e5-logs\") pod \"nova-api-0\" (UID: \"585be2c9-64da-4716-9992-00e1c28380e5\") " pod="openstack/nova-api-0" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.505047 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/585be2c9-64da-4716-9992-00e1c28380e5-public-tls-certs\") pod \"nova-api-0\" (UID: \"585be2c9-64da-4716-9992-00e1c28380e5\") " pod="openstack/nova-api-0" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.506147 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/585be2c9-64da-4716-9992-00e1c28380e5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"585be2c9-64da-4716-9992-00e1c28380e5\") " pod="openstack/nova-api-0" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.506219 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585be2c9-64da-4716-9992-00e1c28380e5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"585be2c9-64da-4716-9992-00e1c28380e5\") " pod="openstack/nova-api-0" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.506556 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/585be2c9-64da-4716-9992-00e1c28380e5-config-data\") pod \"nova-api-0\" (UID: \"585be2c9-64da-4716-9992-00e1c28380e5\") " pod="openstack/nova-api-0" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.527989 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58f8m\" (UniqueName: \"kubernetes.io/projected/585be2c9-64da-4716-9992-00e1c28380e5-kube-api-access-58f8m\") pod \"nova-api-0\" (UID: \"585be2c9-64da-4716-9992-00e1c28380e5\") " pod="openstack/nova-api-0" Oct 02 10:10:11 crc kubenswrapper[4934]: I1002 10:10:11.615826 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:10:12 crc kubenswrapper[4934]: I1002 10:10:12.078560 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:10:12 crc kubenswrapper[4934]: I1002 10:10:12.247926 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6937abdf-169a-4562-8bb4-e54204c7ce14","Type":"ContainerStarted","Data":"d9199437bad8ca6a76f5b2ca7fee5c2e6f5daf028b3950e619bd08a67278229c"} Oct 02 10:10:12 crc kubenswrapper[4934]: I1002 10:10:12.249771 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"585be2c9-64da-4716-9992-00e1c28380e5","Type":"ContainerStarted","Data":"e4f87d31c9d3e55c0735ee3386e293dc29be35d7ce6fd7d6c46b4eedd977ac70"} Oct 02 10:10:12 crc kubenswrapper[4934]: I1002 10:10:12.791122 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:10:12 crc kubenswrapper[4934]: I1002 10:10:12.809992 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:10:12 crc kubenswrapper[4934]: I1002 10:10:12.924992 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6475956e-9f6b-4af0-a1fd-32dc920aae5f" path="/var/lib/kubelet/pods/6475956e-9f6b-4af0-a1fd-32dc920aae5f/volumes" Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.268653 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6937abdf-169a-4562-8bb4-e54204c7ce14","Type":"ContainerStarted","Data":"affe02ef22ab1b1e94a4699a30cc9922a50b1ca924f8918f4b67a079bdc5169b"} Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.268763 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6937abdf-169a-4562-8bb4-e54204c7ce14" containerName="ceilometer-central-agent" containerID="cri-o://8af720ba645d7f4ba050ad0a52b9c6e4c81b6e54491755c67365693a836d27a2" gracePeriod=30 Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.268850 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6937abdf-169a-4562-8bb4-e54204c7ce14" containerName="proxy-httpd" containerID="cri-o://affe02ef22ab1b1e94a4699a30cc9922a50b1ca924f8918f4b67a079bdc5169b" gracePeriod=30 Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.268878 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6937abdf-169a-4562-8bb4-e54204c7ce14" containerName="ceilometer-notification-agent" containerID="cri-o://da90aa2914849d19265600ca300f389c7251f44f36c42ef0c3d0178ade6962d2" gracePeriod=30 Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.268862 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6937abdf-169a-4562-8bb4-e54204c7ce14" containerName="sg-core" containerID="cri-o://d9199437bad8ca6a76f5b2ca7fee5c2e6f5daf028b3950e619bd08a67278229c" gracePeriod=30 Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.269914 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.277911 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"585be2c9-64da-4716-9992-00e1c28380e5","Type":"ContainerStarted","Data":"56c15d17fc08c587b0472bc5d891bd534cfcdd556ee5bd5ae3750ee8dd3c8529"} Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.277970 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"585be2c9-64da-4716-9992-00e1c28380e5","Type":"ContainerStarted","Data":"64f3c0c71c1d78ed7ca72c09a9c0b714710b249faec05fe75efdacefc4e67628"} Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.297991 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.852284882 podStartE2EDuration="5.297974024s" podCreationTimestamp="2025-10-02 10:10:08 +0000 UTC" firstStartedPulling="2025-10-02 10:10:09.08714668 +0000 UTC m=+1280.839788202" lastFinishedPulling="2025-10-02 10:10:12.532835802 +0000 UTC m=+1284.285477344" observedRunningTime="2025-10-02 10:10:13.293296475 +0000 UTC m=+1285.045937997" watchObservedRunningTime="2025-10-02 10:10:13.297974024 +0000 UTC m=+1285.050615546" Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.303257 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.317132 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.317113403 podStartE2EDuration="2.317113403s" podCreationTimestamp="2025-10-02 10:10:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:10:13.310338725 +0000 UTC m=+1285.062980247" watchObservedRunningTime="2025-10-02 10:10:13.317113403 +0000 UTC m=+1285.069754935" Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.498043 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-ldwrv"] Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.499494 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-ldwrv" Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.502354 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.503141 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.507670 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-ldwrv"] Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.535148 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzbv6\" (UniqueName: \"kubernetes.io/projected/4a357498-07a9-439b-9086-b431f165ee7e-kube-api-access-qzbv6\") pod \"nova-cell1-cell-mapping-ldwrv\" (UID: \"4a357498-07a9-439b-9086-b431f165ee7e\") " pod="openstack/nova-cell1-cell-mapping-ldwrv" Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.535197 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a357498-07a9-439b-9086-b431f165ee7e-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-ldwrv\" (UID: \"4a357498-07a9-439b-9086-b431f165ee7e\") " pod="openstack/nova-cell1-cell-mapping-ldwrv" Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.535227 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a357498-07a9-439b-9086-b431f165ee7e-scripts\") pod \"nova-cell1-cell-mapping-ldwrv\" (UID: \"4a357498-07a9-439b-9086-b431f165ee7e\") " pod="openstack/nova-cell1-cell-mapping-ldwrv" Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.535321 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a357498-07a9-439b-9086-b431f165ee7e-config-data\") pod \"nova-cell1-cell-mapping-ldwrv\" (UID: \"4a357498-07a9-439b-9086-b431f165ee7e\") " pod="openstack/nova-cell1-cell-mapping-ldwrv" Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.636588 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzbv6\" (UniqueName: \"kubernetes.io/projected/4a357498-07a9-439b-9086-b431f165ee7e-kube-api-access-qzbv6\") pod \"nova-cell1-cell-mapping-ldwrv\" (UID: \"4a357498-07a9-439b-9086-b431f165ee7e\") " pod="openstack/nova-cell1-cell-mapping-ldwrv" Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.636644 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a357498-07a9-439b-9086-b431f165ee7e-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-ldwrv\" (UID: \"4a357498-07a9-439b-9086-b431f165ee7e\") " pod="openstack/nova-cell1-cell-mapping-ldwrv" Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.636679 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a357498-07a9-439b-9086-b431f165ee7e-scripts\") pod \"nova-cell1-cell-mapping-ldwrv\" (UID: \"4a357498-07a9-439b-9086-b431f165ee7e\") " pod="openstack/nova-cell1-cell-mapping-ldwrv" Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.636820 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a357498-07a9-439b-9086-b431f165ee7e-config-data\") pod \"nova-cell1-cell-mapping-ldwrv\" (UID: \"4a357498-07a9-439b-9086-b431f165ee7e\") " pod="openstack/nova-cell1-cell-mapping-ldwrv" Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.642174 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a357498-07a9-439b-9086-b431f165ee7e-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-ldwrv\" (UID: \"4a357498-07a9-439b-9086-b431f165ee7e\") " pod="openstack/nova-cell1-cell-mapping-ldwrv" Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.642820 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a357498-07a9-439b-9086-b431f165ee7e-scripts\") pod \"nova-cell1-cell-mapping-ldwrv\" (UID: \"4a357498-07a9-439b-9086-b431f165ee7e\") " pod="openstack/nova-cell1-cell-mapping-ldwrv" Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.642975 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a357498-07a9-439b-9086-b431f165ee7e-config-data\") pod \"nova-cell1-cell-mapping-ldwrv\" (UID: \"4a357498-07a9-439b-9086-b431f165ee7e\") " pod="openstack/nova-cell1-cell-mapping-ldwrv" Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.653381 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzbv6\" (UniqueName: \"kubernetes.io/projected/4a357498-07a9-439b-9086-b431f165ee7e-kube-api-access-qzbv6\") pod \"nova-cell1-cell-mapping-ldwrv\" (UID: \"4a357498-07a9-439b-9086-b431f165ee7e\") " pod="openstack/nova-cell1-cell-mapping-ldwrv" Oct 02 10:10:13 crc kubenswrapper[4934]: I1002 10:10:13.856106 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-ldwrv" Oct 02 10:10:14 crc kubenswrapper[4934]: I1002 10:10:14.288594 4934 generic.go:334] "Generic (PLEG): container finished" podID="6937abdf-169a-4562-8bb4-e54204c7ce14" containerID="affe02ef22ab1b1e94a4699a30cc9922a50b1ca924f8918f4b67a079bdc5169b" exitCode=0 Oct 02 10:10:14 crc kubenswrapper[4934]: I1002 10:10:14.288982 4934 generic.go:334] "Generic (PLEG): container finished" podID="6937abdf-169a-4562-8bb4-e54204c7ce14" containerID="d9199437bad8ca6a76f5b2ca7fee5c2e6f5daf028b3950e619bd08a67278229c" exitCode=2 Oct 02 10:10:14 crc kubenswrapper[4934]: I1002 10:10:14.289001 4934 generic.go:334] "Generic (PLEG): container finished" podID="6937abdf-169a-4562-8bb4-e54204c7ce14" containerID="da90aa2914849d19265600ca300f389c7251f44f36c42ef0c3d0178ade6962d2" exitCode=0 Oct 02 10:10:14 crc kubenswrapper[4934]: I1002 10:10:14.288612 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6937abdf-169a-4562-8bb4-e54204c7ce14","Type":"ContainerDied","Data":"affe02ef22ab1b1e94a4699a30cc9922a50b1ca924f8918f4b67a079bdc5169b"} Oct 02 10:10:14 crc kubenswrapper[4934]: I1002 10:10:14.289085 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6937abdf-169a-4562-8bb4-e54204c7ce14","Type":"ContainerDied","Data":"d9199437bad8ca6a76f5b2ca7fee5c2e6f5daf028b3950e619bd08a67278229c"} Oct 02 10:10:14 crc kubenswrapper[4934]: I1002 10:10:14.289099 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6937abdf-169a-4562-8bb4-e54204c7ce14","Type":"ContainerDied","Data":"da90aa2914849d19265600ca300f389c7251f44f36c42ef0c3d0178ade6962d2"} Oct 02 10:10:14 crc kubenswrapper[4934]: I1002 10:10:14.695450 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" Oct 02 10:10:14 crc kubenswrapper[4934]: I1002 10:10:14.766300 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b5cc5d6f9-v9lw6"] Oct 02 10:10:14 crc kubenswrapper[4934]: I1002 10:10:14.766542 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" podUID="a446ab5e-52d1-4962-a383-3fab4a5791f7" containerName="dnsmasq-dns" containerID="cri-o://87ed1325186219049d1c236d9ed445e9035d32d61ebb8891b0fff2e50ae2eb96" gracePeriod=10 Oct 02 10:10:14 crc kubenswrapper[4934]: I1002 10:10:14.958705 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-ldwrv"] Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.292033 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.300636 4934 generic.go:334] "Generic (PLEG): container finished" podID="6937abdf-169a-4562-8bb4-e54204c7ce14" containerID="8af720ba645d7f4ba050ad0a52b9c6e4c81b6e54491755c67365693a836d27a2" exitCode=0 Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.300753 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6937abdf-169a-4562-8bb4-e54204c7ce14","Type":"ContainerDied","Data":"8af720ba645d7f4ba050ad0a52b9c6e4c81b6e54491755c67365693a836d27a2"} Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.302151 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-ldwrv" event={"ID":"4a357498-07a9-439b-9086-b431f165ee7e","Type":"ContainerStarted","Data":"7bf7515a6d0941fe4d48699791654846c0aa70856a9eb0c70058b67cd59fe0cf"} Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.302189 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-ldwrv" event={"ID":"4a357498-07a9-439b-9086-b431f165ee7e","Type":"ContainerStarted","Data":"a7b525a82562e1093704077cfd54fd0a6386e6694e68f09c73452689603faa6c"} Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.303618 4934 generic.go:334] "Generic (PLEG): container finished" podID="a446ab5e-52d1-4962-a383-3fab4a5791f7" containerID="87ed1325186219049d1c236d9ed445e9035d32d61ebb8891b0fff2e50ae2eb96" exitCode=0 Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.303647 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" event={"ID":"a446ab5e-52d1-4962-a383-3fab4a5791f7","Type":"ContainerDied","Data":"87ed1325186219049d1c236d9ed445e9035d32d61ebb8891b0fff2e50ae2eb96"} Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.303677 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" event={"ID":"a446ab5e-52d1-4962-a383-3fab4a5791f7","Type":"ContainerDied","Data":"fa8a18ff773cb7d961fb61e8ce3bd4c70f596c8aefd2d332d278a31ca1093c06"} Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.303694 4934 scope.go:117] "RemoveContainer" containerID="87ed1325186219049d1c236d9ed445e9035d32d61ebb8891b0fff2e50ae2eb96" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.303801 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.312049 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.327421 4934 scope.go:117] "RemoveContainer" containerID="9bbf64eacc67029ce319bc27321ff52fbff4672a4c87120bb01bfa35df7f1b98" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.334655 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-ldwrv" podStartSLOduration=2.334620368 podStartE2EDuration="2.334620368s" podCreationTimestamp="2025-10-02 10:10:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:10:15.331489872 +0000 UTC m=+1287.084131394" watchObservedRunningTime="2025-10-02 10:10:15.334620368 +0000 UTC m=+1287.087261890" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.365833 4934 scope.go:117] "RemoveContainer" containerID="87ed1325186219049d1c236d9ed445e9035d32d61ebb8891b0fff2e50ae2eb96" Oct 02 10:10:15 crc kubenswrapper[4934]: E1002 10:10:15.366324 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87ed1325186219049d1c236d9ed445e9035d32d61ebb8891b0fff2e50ae2eb96\": container with ID starting with 87ed1325186219049d1c236d9ed445e9035d32d61ebb8891b0fff2e50ae2eb96 not found: ID does not exist" containerID="87ed1325186219049d1c236d9ed445e9035d32d61ebb8891b0fff2e50ae2eb96" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.366356 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87ed1325186219049d1c236d9ed445e9035d32d61ebb8891b0fff2e50ae2eb96"} err="failed to get container status \"87ed1325186219049d1c236d9ed445e9035d32d61ebb8891b0fff2e50ae2eb96\": rpc error: code = NotFound desc = could not find container \"87ed1325186219049d1c236d9ed445e9035d32d61ebb8891b0fff2e50ae2eb96\": container with ID starting with 87ed1325186219049d1c236d9ed445e9035d32d61ebb8891b0fff2e50ae2eb96 not found: ID does not exist" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.366376 4934 scope.go:117] "RemoveContainer" containerID="9bbf64eacc67029ce319bc27321ff52fbff4672a4c87120bb01bfa35df7f1b98" Oct 02 10:10:15 crc kubenswrapper[4934]: E1002 10:10:15.370074 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bbf64eacc67029ce319bc27321ff52fbff4672a4c87120bb01bfa35df7f1b98\": container with ID starting with 9bbf64eacc67029ce319bc27321ff52fbff4672a4c87120bb01bfa35df7f1b98 not found: ID does not exist" containerID="9bbf64eacc67029ce319bc27321ff52fbff4672a4c87120bb01bfa35df7f1b98" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.370111 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bbf64eacc67029ce319bc27321ff52fbff4672a4c87120bb01bfa35df7f1b98"} err="failed to get container status \"9bbf64eacc67029ce319bc27321ff52fbff4672a4c87120bb01bfa35df7f1b98\": rpc error: code = NotFound desc = could not find container \"9bbf64eacc67029ce319bc27321ff52fbff4672a4c87120bb01bfa35df7f1b98\": container with ID starting with 9bbf64eacc67029ce319bc27321ff52fbff4672a4c87120bb01bfa35df7f1b98 not found: ID does not exist" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.380799 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-config-data\") pod \"6937abdf-169a-4562-8bb4-e54204c7ce14\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.380854 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-ovsdbserver-nb\") pod \"a446ab5e-52d1-4962-a383-3fab4a5791f7\" (UID: \"a446ab5e-52d1-4962-a383-3fab4a5791f7\") " Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.380902 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-sg-core-conf-yaml\") pod \"6937abdf-169a-4562-8bb4-e54204c7ce14\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.380948 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhxjk\" (UniqueName: \"kubernetes.io/projected/a446ab5e-52d1-4962-a383-3fab4a5791f7-kube-api-access-dhxjk\") pod \"a446ab5e-52d1-4962-a383-3fab4a5791f7\" (UID: \"a446ab5e-52d1-4962-a383-3fab4a5791f7\") " Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.380973 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-dns-swift-storage-0\") pod \"a446ab5e-52d1-4962-a383-3fab4a5791f7\" (UID: \"a446ab5e-52d1-4962-a383-3fab4a5791f7\") " Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.381003 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-config\") pod \"a446ab5e-52d1-4962-a383-3fab4a5791f7\" (UID: \"a446ab5e-52d1-4962-a383-3fab4a5791f7\") " Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.381055 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-combined-ca-bundle\") pod \"6937abdf-169a-4562-8bb4-e54204c7ce14\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.381110 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mgm8\" (UniqueName: \"kubernetes.io/projected/6937abdf-169a-4562-8bb4-e54204c7ce14-kube-api-access-2mgm8\") pod \"6937abdf-169a-4562-8bb4-e54204c7ce14\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.381148 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6937abdf-169a-4562-8bb4-e54204c7ce14-run-httpd\") pod \"6937abdf-169a-4562-8bb4-e54204c7ce14\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.381193 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6937abdf-169a-4562-8bb4-e54204c7ce14-log-httpd\") pod \"6937abdf-169a-4562-8bb4-e54204c7ce14\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.381261 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-ovsdbserver-sb\") pod \"a446ab5e-52d1-4962-a383-3fab4a5791f7\" (UID: \"a446ab5e-52d1-4962-a383-3fab4a5791f7\") " Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.381296 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-scripts\") pod \"6937abdf-169a-4562-8bb4-e54204c7ce14\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.381348 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-ceilometer-tls-certs\") pod \"6937abdf-169a-4562-8bb4-e54204c7ce14\" (UID: \"6937abdf-169a-4562-8bb4-e54204c7ce14\") " Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.381377 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-dns-svc\") pod \"a446ab5e-52d1-4962-a383-3fab4a5791f7\" (UID: \"a446ab5e-52d1-4962-a383-3fab4a5791f7\") " Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.382998 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6937abdf-169a-4562-8bb4-e54204c7ce14-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6937abdf-169a-4562-8bb4-e54204c7ce14" (UID: "6937abdf-169a-4562-8bb4-e54204c7ce14"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.383941 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6937abdf-169a-4562-8bb4-e54204c7ce14-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6937abdf-169a-4562-8bb4-e54204c7ce14" (UID: "6937abdf-169a-4562-8bb4-e54204c7ce14"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.388003 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-scripts" (OuterVolumeSpecName: "scripts") pod "6937abdf-169a-4562-8bb4-e54204c7ce14" (UID: "6937abdf-169a-4562-8bb4-e54204c7ce14"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.388179 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6937abdf-169a-4562-8bb4-e54204c7ce14-kube-api-access-2mgm8" (OuterVolumeSpecName: "kube-api-access-2mgm8") pod "6937abdf-169a-4562-8bb4-e54204c7ce14" (UID: "6937abdf-169a-4562-8bb4-e54204c7ce14"). InnerVolumeSpecName "kube-api-access-2mgm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.398795 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a446ab5e-52d1-4962-a383-3fab4a5791f7-kube-api-access-dhxjk" (OuterVolumeSpecName: "kube-api-access-dhxjk") pod "a446ab5e-52d1-4962-a383-3fab4a5791f7" (UID: "a446ab5e-52d1-4962-a383-3fab4a5791f7"). InnerVolumeSpecName "kube-api-access-dhxjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.420064 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "6937abdf-169a-4562-8bb4-e54204c7ce14" (UID: "6937abdf-169a-4562-8bb4-e54204c7ce14"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.444225 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a446ab5e-52d1-4962-a383-3fab4a5791f7" (UID: "a446ab5e-52d1-4962-a383-3fab4a5791f7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.453273 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a446ab5e-52d1-4962-a383-3fab4a5791f7" (UID: "a446ab5e-52d1-4962-a383-3fab4a5791f7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.454362 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a446ab5e-52d1-4962-a383-3fab4a5791f7" (UID: "a446ab5e-52d1-4962-a383-3fab4a5791f7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.456224 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "6937abdf-169a-4562-8bb4-e54204c7ce14" (UID: "6937abdf-169a-4562-8bb4-e54204c7ce14"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.458795 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-config" (OuterVolumeSpecName: "config") pod "a446ab5e-52d1-4962-a383-3fab4a5791f7" (UID: "a446ab5e-52d1-4962-a383-3fab4a5791f7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.466163 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a446ab5e-52d1-4962-a383-3fab4a5791f7" (UID: "a446ab5e-52d1-4962-a383-3fab4a5791f7"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.475734 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6937abdf-169a-4562-8bb4-e54204c7ce14" (UID: "6937abdf-169a-4562-8bb4-e54204c7ce14"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.483385 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mgm8\" (UniqueName: \"kubernetes.io/projected/6937abdf-169a-4562-8bb4-e54204c7ce14-kube-api-access-2mgm8\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.483416 4934 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6937abdf-169a-4562-8bb4-e54204c7ce14-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.483426 4934 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6937abdf-169a-4562-8bb4-e54204c7ce14-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.483436 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.483444 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.483454 4934 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.483463 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.483471 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.483480 4934 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.483489 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhxjk\" (UniqueName: \"kubernetes.io/projected/a446ab5e-52d1-4962-a383-3fab4a5791f7-kube-api-access-dhxjk\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.483498 4934 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.483507 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a446ab5e-52d1-4962-a383-3fab4a5791f7-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.483515 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.515686 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-config-data" (OuterVolumeSpecName: "config-data") pod "6937abdf-169a-4562-8bb4-e54204c7ce14" (UID: "6937abdf-169a-4562-8bb4-e54204c7ce14"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.584835 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6937abdf-169a-4562-8bb4-e54204c7ce14-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.644062 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b5cc5d6f9-v9lw6"] Oct 02 10:10:15 crc kubenswrapper[4934]: I1002 10:10:15.655230 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b5cc5d6f9-v9lw6"] Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.324372 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6937abdf-169a-4562-8bb4-e54204c7ce14","Type":"ContainerDied","Data":"9b5db17d56e63b66861f15357b776757240748e143fef0ce789beacc510d0d9b"} Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.324661 4934 scope.go:117] "RemoveContainer" containerID="affe02ef22ab1b1e94a4699a30cc9922a50b1ca924f8918f4b67a079bdc5169b" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.324446 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.362178 4934 scope.go:117] "RemoveContainer" containerID="d9199437bad8ca6a76f5b2ca7fee5c2e6f5daf028b3950e619bd08a67278229c" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.367233 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.381013 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.391844 4934 scope.go:117] "RemoveContainer" containerID="da90aa2914849d19265600ca300f389c7251f44f36c42ef0c3d0178ade6962d2" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.395559 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:10:16 crc kubenswrapper[4934]: E1002 10:10:16.396059 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6937abdf-169a-4562-8bb4-e54204c7ce14" containerName="ceilometer-central-agent" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.396080 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="6937abdf-169a-4562-8bb4-e54204c7ce14" containerName="ceilometer-central-agent" Oct 02 10:10:16 crc kubenswrapper[4934]: E1002 10:10:16.396110 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a446ab5e-52d1-4962-a383-3fab4a5791f7" containerName="init" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.396121 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a446ab5e-52d1-4962-a383-3fab4a5791f7" containerName="init" Oct 02 10:10:16 crc kubenswrapper[4934]: E1002 10:10:16.396142 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6937abdf-169a-4562-8bb4-e54204c7ce14" containerName="ceilometer-notification-agent" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.396152 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="6937abdf-169a-4562-8bb4-e54204c7ce14" containerName="ceilometer-notification-agent" Oct 02 10:10:16 crc kubenswrapper[4934]: E1002 10:10:16.396167 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6937abdf-169a-4562-8bb4-e54204c7ce14" containerName="proxy-httpd" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.396176 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="6937abdf-169a-4562-8bb4-e54204c7ce14" containerName="proxy-httpd" Oct 02 10:10:16 crc kubenswrapper[4934]: E1002 10:10:16.396192 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a446ab5e-52d1-4962-a383-3fab4a5791f7" containerName="dnsmasq-dns" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.396202 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a446ab5e-52d1-4962-a383-3fab4a5791f7" containerName="dnsmasq-dns" Oct 02 10:10:16 crc kubenswrapper[4934]: E1002 10:10:16.396219 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6937abdf-169a-4562-8bb4-e54204c7ce14" containerName="sg-core" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.396227 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="6937abdf-169a-4562-8bb4-e54204c7ce14" containerName="sg-core" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.396472 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="6937abdf-169a-4562-8bb4-e54204c7ce14" containerName="proxy-httpd" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.396497 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="6937abdf-169a-4562-8bb4-e54204c7ce14" containerName="ceilometer-central-agent" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.396509 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="a446ab5e-52d1-4962-a383-3fab4a5791f7" containerName="dnsmasq-dns" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.396521 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="6937abdf-169a-4562-8bb4-e54204c7ce14" containerName="sg-core" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.396537 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="6937abdf-169a-4562-8bb4-e54204c7ce14" containerName="ceilometer-notification-agent" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.408871 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.411045 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.411996 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.413481 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.413924 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.425763 4934 scope.go:117] "RemoveContainer" containerID="8af720ba645d7f4ba050ad0a52b9c6e4c81b6e54491755c67365693a836d27a2" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.500476 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4becb20a-8bf6-458d-856e-c8d6cd43f797-log-httpd\") pod \"ceilometer-0\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " pod="openstack/ceilometer-0" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.501256 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-scripts\") pod \"ceilometer-0\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " pod="openstack/ceilometer-0" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.501350 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-config-data\") pod \"ceilometer-0\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " pod="openstack/ceilometer-0" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.501487 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6z6tn\" (UniqueName: \"kubernetes.io/projected/4becb20a-8bf6-458d-856e-c8d6cd43f797-kube-api-access-6z6tn\") pod \"ceilometer-0\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " pod="openstack/ceilometer-0" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.501591 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4becb20a-8bf6-458d-856e-c8d6cd43f797-run-httpd\") pod \"ceilometer-0\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " pod="openstack/ceilometer-0" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.501719 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " pod="openstack/ceilometer-0" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.501803 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " pod="openstack/ceilometer-0" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.501866 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " pod="openstack/ceilometer-0" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.603831 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4becb20a-8bf6-458d-856e-c8d6cd43f797-run-httpd\") pod \"ceilometer-0\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " pod="openstack/ceilometer-0" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.604144 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " pod="openstack/ceilometer-0" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.605145 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " pod="openstack/ceilometer-0" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.605242 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " pod="openstack/ceilometer-0" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.605338 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4becb20a-8bf6-458d-856e-c8d6cd43f797-log-httpd\") pod \"ceilometer-0\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " pod="openstack/ceilometer-0" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.604347 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4becb20a-8bf6-458d-856e-c8d6cd43f797-run-httpd\") pod \"ceilometer-0\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " pod="openstack/ceilometer-0" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.605471 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-scripts\") pod \"ceilometer-0\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " pod="openstack/ceilometer-0" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.605702 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-config-data\") pod \"ceilometer-0\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " pod="openstack/ceilometer-0" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.605843 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6z6tn\" (UniqueName: \"kubernetes.io/projected/4becb20a-8bf6-458d-856e-c8d6cd43f797-kube-api-access-6z6tn\") pod \"ceilometer-0\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " pod="openstack/ceilometer-0" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.605929 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4becb20a-8bf6-458d-856e-c8d6cd43f797-log-httpd\") pod \"ceilometer-0\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " pod="openstack/ceilometer-0" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.610048 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-scripts\") pod \"ceilometer-0\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " pod="openstack/ceilometer-0" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.612208 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " pod="openstack/ceilometer-0" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.612301 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " pod="openstack/ceilometer-0" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.609872 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " pod="openstack/ceilometer-0" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.613772 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-config-data\") pod \"ceilometer-0\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " pod="openstack/ceilometer-0" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.627769 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6z6tn\" (UniqueName: \"kubernetes.io/projected/4becb20a-8bf6-458d-856e-c8d6cd43f797-kube-api-access-6z6tn\") pod \"ceilometer-0\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " pod="openstack/ceilometer-0" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.743786 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.924198 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6937abdf-169a-4562-8bb4-e54204c7ce14" path="/var/lib/kubelet/pods/6937abdf-169a-4562-8bb4-e54204c7ce14/volumes" Oct 02 10:10:16 crc kubenswrapper[4934]: I1002 10:10:16.925232 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a446ab5e-52d1-4962-a383-3fab4a5791f7" path="/var/lib/kubelet/pods/a446ab5e-52d1-4962-a383-3fab4a5791f7/volumes" Oct 02 10:10:17 crc kubenswrapper[4934]: I1002 10:10:17.256498 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:10:17 crc kubenswrapper[4934]: I1002 10:10:17.268400 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 10:10:17 crc kubenswrapper[4934]: I1002 10:10:17.346628 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4becb20a-8bf6-458d-856e-c8d6cd43f797","Type":"ContainerStarted","Data":"01e10a448cab604b193449c027d3bef612ac49c44b1079e5df9e7dbe2ca1bfd1"} Oct 02 10:10:18 crc kubenswrapper[4934]: I1002 10:10:18.367562 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4becb20a-8bf6-458d-856e-c8d6cd43f797","Type":"ContainerStarted","Data":"bf2e8afd472c84d3385087f6698782dc92f0a62388c560ef85d4823b8156cfb7"} Oct 02 10:10:19 crc kubenswrapper[4934]: I1002 10:10:19.382488 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4becb20a-8bf6-458d-856e-c8d6cd43f797","Type":"ContainerStarted","Data":"8e5fcc5a4c69ba03bdc17a733d109d36416615950178c66556ce847df4a0f0cf"} Oct 02 10:10:20 crc kubenswrapper[4934]: I1002 10:10:20.269455 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b5cc5d6f9-v9lw6" podUID="a446ab5e-52d1-4962-a383-3fab4a5791f7" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.189:5353: i/o timeout" Oct 02 10:10:20 crc kubenswrapper[4934]: I1002 10:10:20.392720 4934 generic.go:334] "Generic (PLEG): container finished" podID="4a357498-07a9-439b-9086-b431f165ee7e" containerID="7bf7515a6d0941fe4d48699791654846c0aa70856a9eb0c70058b67cd59fe0cf" exitCode=0 Oct 02 10:10:20 crc kubenswrapper[4934]: I1002 10:10:20.392786 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-ldwrv" event={"ID":"4a357498-07a9-439b-9086-b431f165ee7e","Type":"ContainerDied","Data":"7bf7515a6d0941fe4d48699791654846c0aa70856a9eb0c70058b67cd59fe0cf"} Oct 02 10:10:20 crc kubenswrapper[4934]: I1002 10:10:20.395221 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4becb20a-8bf6-458d-856e-c8d6cd43f797","Type":"ContainerStarted","Data":"1b462ec2465f66c7ac2fdf185d407ece6092286b180138bdcd246fc6359029b3"} Oct 02 10:10:21 crc kubenswrapper[4934]: I1002 10:10:21.415713 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4becb20a-8bf6-458d-856e-c8d6cd43f797","Type":"ContainerStarted","Data":"8cdeead710c71e753ff7516bcb319070ccf923a448f85b9b738b5bbc64e5abd1"} Oct 02 10:10:21 crc kubenswrapper[4934]: I1002 10:10:21.416194 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 10:10:21 crc kubenswrapper[4934]: I1002 10:10:21.480326 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.979949075 podStartE2EDuration="5.480305166s" podCreationTimestamp="2025-10-02 10:10:16 +0000 UTC" firstStartedPulling="2025-10-02 10:10:17.268053032 +0000 UTC m=+1289.020694564" lastFinishedPulling="2025-10-02 10:10:20.768409133 +0000 UTC m=+1292.521050655" observedRunningTime="2025-10-02 10:10:21.456474297 +0000 UTC m=+1293.209115839" watchObservedRunningTime="2025-10-02 10:10:21.480305166 +0000 UTC m=+1293.232946698" Oct 02 10:10:21 crc kubenswrapper[4934]: I1002 10:10:21.616868 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 10:10:21 crc kubenswrapper[4934]: I1002 10:10:21.616927 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 10:10:21 crc kubenswrapper[4934]: I1002 10:10:21.838740 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-ldwrv" Oct 02 10:10:21 crc kubenswrapper[4934]: I1002 10:10:21.913117 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a357498-07a9-439b-9086-b431f165ee7e-combined-ca-bundle\") pod \"4a357498-07a9-439b-9086-b431f165ee7e\" (UID: \"4a357498-07a9-439b-9086-b431f165ee7e\") " Oct 02 10:10:21 crc kubenswrapper[4934]: I1002 10:10:21.913347 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a357498-07a9-439b-9086-b431f165ee7e-config-data\") pod \"4a357498-07a9-439b-9086-b431f165ee7e\" (UID: \"4a357498-07a9-439b-9086-b431f165ee7e\") " Oct 02 10:10:21 crc kubenswrapper[4934]: I1002 10:10:21.913387 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a357498-07a9-439b-9086-b431f165ee7e-scripts\") pod \"4a357498-07a9-439b-9086-b431f165ee7e\" (UID: \"4a357498-07a9-439b-9086-b431f165ee7e\") " Oct 02 10:10:21 crc kubenswrapper[4934]: I1002 10:10:21.913454 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzbv6\" (UniqueName: \"kubernetes.io/projected/4a357498-07a9-439b-9086-b431f165ee7e-kube-api-access-qzbv6\") pod \"4a357498-07a9-439b-9086-b431f165ee7e\" (UID: \"4a357498-07a9-439b-9086-b431f165ee7e\") " Oct 02 10:10:21 crc kubenswrapper[4934]: I1002 10:10:21.919668 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a357498-07a9-439b-9086-b431f165ee7e-scripts" (OuterVolumeSpecName: "scripts") pod "4a357498-07a9-439b-9086-b431f165ee7e" (UID: "4a357498-07a9-439b-9086-b431f165ee7e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:21 crc kubenswrapper[4934]: I1002 10:10:21.929475 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a357498-07a9-439b-9086-b431f165ee7e-kube-api-access-qzbv6" (OuterVolumeSpecName: "kube-api-access-qzbv6") pod "4a357498-07a9-439b-9086-b431f165ee7e" (UID: "4a357498-07a9-439b-9086-b431f165ee7e"). InnerVolumeSpecName "kube-api-access-qzbv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:10:21 crc kubenswrapper[4934]: I1002 10:10:21.958777 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a357498-07a9-439b-9086-b431f165ee7e-config-data" (OuterVolumeSpecName: "config-data") pod "4a357498-07a9-439b-9086-b431f165ee7e" (UID: "4a357498-07a9-439b-9086-b431f165ee7e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:21 crc kubenswrapper[4934]: I1002 10:10:21.964389 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a357498-07a9-439b-9086-b431f165ee7e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a357498-07a9-439b-9086-b431f165ee7e" (UID: "4a357498-07a9-439b-9086-b431f165ee7e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:22 crc kubenswrapper[4934]: I1002 10:10:22.015287 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a357498-07a9-439b-9086-b431f165ee7e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:22 crc kubenswrapper[4934]: I1002 10:10:22.015321 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a357498-07a9-439b-9086-b431f165ee7e-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:22 crc kubenswrapper[4934]: I1002 10:10:22.015331 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a357498-07a9-439b-9086-b431f165ee7e-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:22 crc kubenswrapper[4934]: I1002 10:10:22.015340 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzbv6\" (UniqueName: \"kubernetes.io/projected/4a357498-07a9-439b-9086-b431f165ee7e-kube-api-access-qzbv6\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:22 crc kubenswrapper[4934]: I1002 10:10:22.424136 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-ldwrv" event={"ID":"4a357498-07a9-439b-9086-b431f165ee7e","Type":"ContainerDied","Data":"a7b525a82562e1093704077cfd54fd0a6386e6694e68f09c73452689603faa6c"} Oct 02 10:10:22 crc kubenswrapper[4934]: I1002 10:10:22.424191 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7b525a82562e1093704077cfd54fd0a6386e6694e68f09c73452689603faa6c" Oct 02 10:10:22 crc kubenswrapper[4934]: I1002 10:10:22.424160 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-ldwrv" Oct 02 10:10:22 crc kubenswrapper[4934]: I1002 10:10:22.628747 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="585be2c9-64da-4716-9992-00e1c28380e5" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.200:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 10:10:22 crc kubenswrapper[4934]: I1002 10:10:22.628770 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="585be2c9-64da-4716-9992-00e1c28380e5" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.200:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 10:10:22 crc kubenswrapper[4934]: I1002 10:10:22.645286 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:10:22 crc kubenswrapper[4934]: I1002 10:10:22.645510 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="585be2c9-64da-4716-9992-00e1c28380e5" containerName="nova-api-log" containerID="cri-o://64f3c0c71c1d78ed7ca72c09a9c0b714710b249faec05fe75efdacefc4e67628" gracePeriod=30 Oct 02 10:10:22 crc kubenswrapper[4934]: I1002 10:10:22.645570 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="585be2c9-64da-4716-9992-00e1c28380e5" containerName="nova-api-api" containerID="cri-o://56c15d17fc08c587b0472bc5d891bd534cfcdd556ee5bd5ae3750ee8dd3c8529" gracePeriod=30 Oct 02 10:10:22 crc kubenswrapper[4934]: I1002 10:10:22.661562 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:10:22 crc kubenswrapper[4934]: I1002 10:10:22.661912 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="eeb3879e-b8fc-43ab-ad26-42ae80715706" containerName="nova-scheduler-scheduler" containerID="cri-o://c6b414c90f3b2c8ee4f300f9213c49d4f93007c4ebe815709d38d734fde7cf3b" gracePeriod=30 Oct 02 10:10:22 crc kubenswrapper[4934]: I1002 10:10:22.721484 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:10:22 crc kubenswrapper[4934]: I1002 10:10:22.721716 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="edb4b618-5f25-4a83-a2e7-3b2e2d6a2049" containerName="nova-metadata-log" containerID="cri-o://de95cb4af5f7758289a9ed99c925c143cf2617fd14807a19a8624ee861cf0c14" gracePeriod=30 Oct 02 10:10:22 crc kubenswrapper[4934]: I1002 10:10:22.721755 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="edb4b618-5f25-4a83-a2e7-3b2e2d6a2049" containerName="nova-metadata-metadata" containerID="cri-o://2d1cfcb32101f27ad01e6c51bb7828a926d69c41b6d7fa00b917776ca6b15cc4" gracePeriod=30 Oct 02 10:10:23 crc kubenswrapper[4934]: I1002 10:10:23.438349 4934 generic.go:334] "Generic (PLEG): container finished" podID="585be2c9-64da-4716-9992-00e1c28380e5" containerID="64f3c0c71c1d78ed7ca72c09a9c0b714710b249faec05fe75efdacefc4e67628" exitCode=143 Oct 02 10:10:23 crc kubenswrapper[4934]: I1002 10:10:23.438431 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"585be2c9-64da-4716-9992-00e1c28380e5","Type":"ContainerDied","Data":"64f3c0c71c1d78ed7ca72c09a9c0b714710b249faec05fe75efdacefc4e67628"} Oct 02 10:10:23 crc kubenswrapper[4934]: I1002 10:10:23.440932 4934 generic.go:334] "Generic (PLEG): container finished" podID="edb4b618-5f25-4a83-a2e7-3b2e2d6a2049" containerID="de95cb4af5f7758289a9ed99c925c143cf2617fd14807a19a8624ee861cf0c14" exitCode=143 Oct 02 10:10:23 crc kubenswrapper[4934]: I1002 10:10:23.441005 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049","Type":"ContainerDied","Data":"de95cb4af5f7758289a9ed99c925c143cf2617fd14807a19a8624ee861cf0c14"} Oct 02 10:10:24 crc kubenswrapper[4934]: E1002 10:10:24.263547 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c6b414c90f3b2c8ee4f300f9213c49d4f93007c4ebe815709d38d734fde7cf3b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 10:10:24 crc kubenswrapper[4934]: E1002 10:10:24.265153 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c6b414c90f3b2c8ee4f300f9213c49d4f93007c4ebe815709d38d734fde7cf3b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 10:10:24 crc kubenswrapper[4934]: E1002 10:10:24.266325 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c6b414c90f3b2c8ee4f300f9213c49d4f93007c4ebe815709d38d734fde7cf3b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 10:10:24 crc kubenswrapper[4934]: E1002 10:10:24.266376 4934 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="eeb3879e-b8fc-43ab-ad26-42ae80715706" containerName="nova-scheduler-scheduler" Oct 02 10:10:25 crc kubenswrapper[4934]: I1002 10:10:25.868562 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="edb4b618-5f25-4a83-a2e7-3b2e2d6a2049" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.192:8775/\": read tcp 10.217.0.2:36826->10.217.0.192:8775: read: connection reset by peer" Oct 02 10:10:25 crc kubenswrapper[4934]: I1002 10:10:25.868563 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="edb4b618-5f25-4a83-a2e7-3b2e2d6a2049" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.192:8775/\": read tcp 10.217.0.2:36832->10.217.0.192:8775: read: connection reset by peer" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.442444 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.511509 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-config-data\") pod \"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049\" (UID: \"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049\") " Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.511891 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-combined-ca-bundle\") pod \"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049\" (UID: \"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049\") " Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.511925 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-nova-metadata-tls-certs\") pod \"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049\" (UID: \"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049\") " Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.511960 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2fpg\" (UniqueName: \"kubernetes.io/projected/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-kube-api-access-p2fpg\") pod \"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049\" (UID: \"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049\") " Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.512004 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-logs\") pod \"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049\" (UID: \"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049\") " Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.513124 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-logs" (OuterVolumeSpecName: "logs") pod "edb4b618-5f25-4a83-a2e7-3b2e2d6a2049" (UID: "edb4b618-5f25-4a83-a2e7-3b2e2d6a2049"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.513350 4934 generic.go:334] "Generic (PLEG): container finished" podID="edb4b618-5f25-4a83-a2e7-3b2e2d6a2049" containerID="2d1cfcb32101f27ad01e6c51bb7828a926d69c41b6d7fa00b917776ca6b15cc4" exitCode=0 Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.513392 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049","Type":"ContainerDied","Data":"2d1cfcb32101f27ad01e6c51bb7828a926d69c41b6d7fa00b917776ca6b15cc4"} Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.513425 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"edb4b618-5f25-4a83-a2e7-3b2e2d6a2049","Type":"ContainerDied","Data":"fcf0e04f0d8939aa100d8c0efadcf5f98a2780516f6570c44a389946090c1ea6"} Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.513447 4934 scope.go:117] "RemoveContainer" containerID="2d1cfcb32101f27ad01e6c51bb7828a926d69c41b6d7fa00b917776ca6b15cc4" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.513659 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.550884 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-kube-api-access-p2fpg" (OuterVolumeSpecName: "kube-api-access-p2fpg") pod "edb4b618-5f25-4a83-a2e7-3b2e2d6a2049" (UID: "edb4b618-5f25-4a83-a2e7-3b2e2d6a2049"). InnerVolumeSpecName "kube-api-access-p2fpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.563645 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "edb4b618-5f25-4a83-a2e7-3b2e2d6a2049" (UID: "edb4b618-5f25-4a83-a2e7-3b2e2d6a2049"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.574738 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-config-data" (OuterVolumeSpecName: "config-data") pod "edb4b618-5f25-4a83-a2e7-3b2e2d6a2049" (UID: "edb4b618-5f25-4a83-a2e7-3b2e2d6a2049"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.604226 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "edb4b618-5f25-4a83-a2e7-3b2e2d6a2049" (UID: "edb4b618-5f25-4a83-a2e7-3b2e2d6a2049"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.615754 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.615781 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.615791 4934 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.615800 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2fpg\" (UniqueName: \"kubernetes.io/projected/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-kube-api-access-p2fpg\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.615809 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.639326 4934 scope.go:117] "RemoveContainer" containerID="de95cb4af5f7758289a9ed99c925c143cf2617fd14807a19a8624ee861cf0c14" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.660660 4934 scope.go:117] "RemoveContainer" containerID="2d1cfcb32101f27ad01e6c51bb7828a926d69c41b6d7fa00b917776ca6b15cc4" Oct 02 10:10:26 crc kubenswrapper[4934]: E1002 10:10:26.661125 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d1cfcb32101f27ad01e6c51bb7828a926d69c41b6d7fa00b917776ca6b15cc4\": container with ID starting with 2d1cfcb32101f27ad01e6c51bb7828a926d69c41b6d7fa00b917776ca6b15cc4 not found: ID does not exist" containerID="2d1cfcb32101f27ad01e6c51bb7828a926d69c41b6d7fa00b917776ca6b15cc4" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.661156 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d1cfcb32101f27ad01e6c51bb7828a926d69c41b6d7fa00b917776ca6b15cc4"} err="failed to get container status \"2d1cfcb32101f27ad01e6c51bb7828a926d69c41b6d7fa00b917776ca6b15cc4\": rpc error: code = NotFound desc = could not find container \"2d1cfcb32101f27ad01e6c51bb7828a926d69c41b6d7fa00b917776ca6b15cc4\": container with ID starting with 2d1cfcb32101f27ad01e6c51bb7828a926d69c41b6d7fa00b917776ca6b15cc4 not found: ID does not exist" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.661179 4934 scope.go:117] "RemoveContainer" containerID="de95cb4af5f7758289a9ed99c925c143cf2617fd14807a19a8624ee861cf0c14" Oct 02 10:10:26 crc kubenswrapper[4934]: E1002 10:10:26.661427 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de95cb4af5f7758289a9ed99c925c143cf2617fd14807a19a8624ee861cf0c14\": container with ID starting with de95cb4af5f7758289a9ed99c925c143cf2617fd14807a19a8624ee861cf0c14 not found: ID does not exist" containerID="de95cb4af5f7758289a9ed99c925c143cf2617fd14807a19a8624ee861cf0c14" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.661449 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de95cb4af5f7758289a9ed99c925c143cf2617fd14807a19a8624ee861cf0c14"} err="failed to get container status \"de95cb4af5f7758289a9ed99c925c143cf2617fd14807a19a8624ee861cf0c14\": rpc error: code = NotFound desc = could not find container \"de95cb4af5f7758289a9ed99c925c143cf2617fd14807a19a8624ee861cf0c14\": container with ID starting with de95cb4af5f7758289a9ed99c925c143cf2617fd14807a19a8624ee861cf0c14 not found: ID does not exist" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.847346 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.855283 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.865921 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:10:26 crc kubenswrapper[4934]: E1002 10:10:26.866264 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edb4b618-5f25-4a83-a2e7-3b2e2d6a2049" containerName="nova-metadata-log" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.866283 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="edb4b618-5f25-4a83-a2e7-3b2e2d6a2049" containerName="nova-metadata-log" Oct 02 10:10:26 crc kubenswrapper[4934]: E1002 10:10:26.866318 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edb4b618-5f25-4a83-a2e7-3b2e2d6a2049" containerName="nova-metadata-metadata" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.866325 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="edb4b618-5f25-4a83-a2e7-3b2e2d6a2049" containerName="nova-metadata-metadata" Oct 02 10:10:26 crc kubenswrapper[4934]: E1002 10:10:26.866333 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a357498-07a9-439b-9086-b431f165ee7e" containerName="nova-manage" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.866339 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a357498-07a9-439b-9086-b431f165ee7e" containerName="nova-manage" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.866515 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="edb4b618-5f25-4a83-a2e7-3b2e2d6a2049" containerName="nova-metadata-log" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.866533 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a357498-07a9-439b-9086-b431f165ee7e" containerName="nova-manage" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.866556 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="edb4b618-5f25-4a83-a2e7-3b2e2d6a2049" containerName="nova-metadata-metadata" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.867547 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.871421 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.871675 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.881532 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.919987 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04aff378-509c-49be-bf4c-2d59111fc910-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"04aff378-509c-49be-bf4c-2d59111fc910\") " pod="openstack/nova-metadata-0" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.920063 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04aff378-509c-49be-bf4c-2d59111fc910-logs\") pod \"nova-metadata-0\" (UID: \"04aff378-509c-49be-bf4c-2d59111fc910\") " pod="openstack/nova-metadata-0" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.920303 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/04aff378-509c-49be-bf4c-2d59111fc910-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"04aff378-509c-49be-bf4c-2d59111fc910\") " pod="openstack/nova-metadata-0" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.920488 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04aff378-509c-49be-bf4c-2d59111fc910-config-data\") pod \"nova-metadata-0\" (UID: \"04aff378-509c-49be-bf4c-2d59111fc910\") " pod="openstack/nova-metadata-0" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.920559 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4spw\" (UniqueName: \"kubernetes.io/projected/04aff378-509c-49be-bf4c-2d59111fc910-kube-api-access-z4spw\") pod \"nova-metadata-0\" (UID: \"04aff378-509c-49be-bf4c-2d59111fc910\") " pod="openstack/nova-metadata-0" Oct 02 10:10:26 crc kubenswrapper[4934]: I1002 10:10:26.923437 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edb4b618-5f25-4a83-a2e7-3b2e2d6a2049" path="/var/lib/kubelet/pods/edb4b618-5f25-4a83-a2e7-3b2e2d6a2049/volumes" Oct 02 10:10:27 crc kubenswrapper[4934]: I1002 10:10:27.021749 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04aff378-509c-49be-bf4c-2d59111fc910-config-data\") pod \"nova-metadata-0\" (UID: \"04aff378-509c-49be-bf4c-2d59111fc910\") " pod="openstack/nova-metadata-0" Oct 02 10:10:27 crc kubenswrapper[4934]: I1002 10:10:27.021805 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4spw\" (UniqueName: \"kubernetes.io/projected/04aff378-509c-49be-bf4c-2d59111fc910-kube-api-access-z4spw\") pod \"nova-metadata-0\" (UID: \"04aff378-509c-49be-bf4c-2d59111fc910\") " pod="openstack/nova-metadata-0" Oct 02 10:10:27 crc kubenswrapper[4934]: I1002 10:10:27.021921 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04aff378-509c-49be-bf4c-2d59111fc910-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"04aff378-509c-49be-bf4c-2d59111fc910\") " pod="openstack/nova-metadata-0" Oct 02 10:10:27 crc kubenswrapper[4934]: I1002 10:10:27.021978 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04aff378-509c-49be-bf4c-2d59111fc910-logs\") pod \"nova-metadata-0\" (UID: \"04aff378-509c-49be-bf4c-2d59111fc910\") " pod="openstack/nova-metadata-0" Oct 02 10:10:27 crc kubenswrapper[4934]: I1002 10:10:27.022055 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/04aff378-509c-49be-bf4c-2d59111fc910-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"04aff378-509c-49be-bf4c-2d59111fc910\") " pod="openstack/nova-metadata-0" Oct 02 10:10:27 crc kubenswrapper[4934]: I1002 10:10:27.022978 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04aff378-509c-49be-bf4c-2d59111fc910-logs\") pod \"nova-metadata-0\" (UID: \"04aff378-509c-49be-bf4c-2d59111fc910\") " pod="openstack/nova-metadata-0" Oct 02 10:10:27 crc kubenswrapper[4934]: I1002 10:10:27.026683 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04aff378-509c-49be-bf4c-2d59111fc910-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"04aff378-509c-49be-bf4c-2d59111fc910\") " pod="openstack/nova-metadata-0" Oct 02 10:10:27 crc kubenswrapper[4934]: I1002 10:10:27.027532 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/04aff378-509c-49be-bf4c-2d59111fc910-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"04aff378-509c-49be-bf4c-2d59111fc910\") " pod="openstack/nova-metadata-0" Oct 02 10:10:27 crc kubenswrapper[4934]: I1002 10:10:27.027678 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04aff378-509c-49be-bf4c-2d59111fc910-config-data\") pod \"nova-metadata-0\" (UID: \"04aff378-509c-49be-bf4c-2d59111fc910\") " pod="openstack/nova-metadata-0" Oct 02 10:10:27 crc kubenswrapper[4934]: I1002 10:10:27.041123 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4spw\" (UniqueName: \"kubernetes.io/projected/04aff378-509c-49be-bf4c-2d59111fc910-kube-api-access-z4spw\") pod \"nova-metadata-0\" (UID: \"04aff378-509c-49be-bf4c-2d59111fc910\") " pod="openstack/nova-metadata-0" Oct 02 10:10:27 crc kubenswrapper[4934]: I1002 10:10:27.183501 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 10:10:27 crc kubenswrapper[4934]: I1002 10:10:27.653967 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:10:27 crc kubenswrapper[4934]: W1002 10:10:27.659372 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04aff378_509c_49be_bf4c_2d59111fc910.slice/crio-d810a75aec31cfdfe545e4469e243df734555ffbf92e9708fa55b94d75895215 WatchSource:0}: Error finding container d810a75aec31cfdfe545e4469e243df734555ffbf92e9708fa55b94d75895215: Status 404 returned error can't find the container with id d810a75aec31cfdfe545e4469e243df734555ffbf92e9708fa55b94d75895215 Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.350073 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.448829 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb3879e-b8fc-43ab-ad26-42ae80715706-combined-ca-bundle\") pod \"eeb3879e-b8fc-43ab-ad26-42ae80715706\" (UID: \"eeb3879e-b8fc-43ab-ad26-42ae80715706\") " Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.449121 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-th89f\" (UniqueName: \"kubernetes.io/projected/eeb3879e-b8fc-43ab-ad26-42ae80715706-kube-api-access-th89f\") pod \"eeb3879e-b8fc-43ab-ad26-42ae80715706\" (UID: \"eeb3879e-b8fc-43ab-ad26-42ae80715706\") " Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.449222 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eeb3879e-b8fc-43ab-ad26-42ae80715706-config-data\") pod \"eeb3879e-b8fc-43ab-ad26-42ae80715706\" (UID: \"eeb3879e-b8fc-43ab-ad26-42ae80715706\") " Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.456219 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eeb3879e-b8fc-43ab-ad26-42ae80715706-kube-api-access-th89f" (OuterVolumeSpecName: "kube-api-access-th89f") pod "eeb3879e-b8fc-43ab-ad26-42ae80715706" (UID: "eeb3879e-b8fc-43ab-ad26-42ae80715706"). InnerVolumeSpecName "kube-api-access-th89f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.478551 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.488940 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eeb3879e-b8fc-43ab-ad26-42ae80715706-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eeb3879e-b8fc-43ab-ad26-42ae80715706" (UID: "eeb3879e-b8fc-43ab-ad26-42ae80715706"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.496811 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eeb3879e-b8fc-43ab-ad26-42ae80715706-config-data" (OuterVolumeSpecName: "config-data") pod "eeb3879e-b8fc-43ab-ad26-42ae80715706" (UID: "eeb3879e-b8fc-43ab-ad26-42ae80715706"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.538089 4934 generic.go:334] "Generic (PLEG): container finished" podID="585be2c9-64da-4716-9992-00e1c28380e5" containerID="56c15d17fc08c587b0472bc5d891bd534cfcdd556ee5bd5ae3750ee8dd3c8529" exitCode=0 Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.538143 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.538188 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"585be2c9-64da-4716-9992-00e1c28380e5","Type":"ContainerDied","Data":"56c15d17fc08c587b0472bc5d891bd534cfcdd556ee5bd5ae3750ee8dd3c8529"} Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.538218 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"585be2c9-64da-4716-9992-00e1c28380e5","Type":"ContainerDied","Data":"e4f87d31c9d3e55c0735ee3386e293dc29be35d7ce6fd7d6c46b4eedd977ac70"} Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.538234 4934 scope.go:117] "RemoveContainer" containerID="56c15d17fc08c587b0472bc5d891bd534cfcdd556ee5bd5ae3750ee8dd3c8529" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.540426 4934 generic.go:334] "Generic (PLEG): container finished" podID="eeb3879e-b8fc-43ab-ad26-42ae80715706" containerID="c6b414c90f3b2c8ee4f300f9213c49d4f93007c4ebe815709d38d734fde7cf3b" exitCode=0 Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.540495 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"eeb3879e-b8fc-43ab-ad26-42ae80715706","Type":"ContainerDied","Data":"c6b414c90f3b2c8ee4f300f9213c49d4f93007c4ebe815709d38d734fde7cf3b"} Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.540526 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"eeb3879e-b8fc-43ab-ad26-42ae80715706","Type":"ContainerDied","Data":"a14bbcf611a2af141e70f921257f49fbe1eb653bd690ef1c6acab5a466d9415d"} Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.540555 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.548800 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"04aff378-509c-49be-bf4c-2d59111fc910","Type":"ContainerStarted","Data":"f28c0b65a711c22b95f0ba26fc6761c9a173fc0e9b32cd123f69c153b87e25bf"} Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.548839 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"04aff378-509c-49be-bf4c-2d59111fc910","Type":"ContainerStarted","Data":"0c87adf381aab6824e14a94acbf03d404203f8ec3721590d4efd316ce48e8dd8"} Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.548850 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"04aff378-509c-49be-bf4c-2d59111fc910","Type":"ContainerStarted","Data":"d810a75aec31cfdfe545e4469e243df734555ffbf92e9708fa55b94d75895215"} Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.550124 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/585be2c9-64da-4716-9992-00e1c28380e5-config-data\") pod \"585be2c9-64da-4716-9992-00e1c28380e5\" (UID: \"585be2c9-64da-4716-9992-00e1c28380e5\") " Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.550147 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/585be2c9-64da-4716-9992-00e1c28380e5-internal-tls-certs\") pod \"585be2c9-64da-4716-9992-00e1c28380e5\" (UID: \"585be2c9-64da-4716-9992-00e1c28380e5\") " Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.550264 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/585be2c9-64da-4716-9992-00e1c28380e5-logs\") pod \"585be2c9-64da-4716-9992-00e1c28380e5\" (UID: \"585be2c9-64da-4716-9992-00e1c28380e5\") " Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.550329 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/585be2c9-64da-4716-9992-00e1c28380e5-public-tls-certs\") pod \"585be2c9-64da-4716-9992-00e1c28380e5\" (UID: \"585be2c9-64da-4716-9992-00e1c28380e5\") " Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.550354 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585be2c9-64da-4716-9992-00e1c28380e5-combined-ca-bundle\") pod \"585be2c9-64da-4716-9992-00e1c28380e5\" (UID: \"585be2c9-64da-4716-9992-00e1c28380e5\") " Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.550460 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58f8m\" (UniqueName: \"kubernetes.io/projected/585be2c9-64da-4716-9992-00e1c28380e5-kube-api-access-58f8m\") pod \"585be2c9-64da-4716-9992-00e1c28380e5\" (UID: \"585be2c9-64da-4716-9992-00e1c28380e5\") " Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.550817 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb3879e-b8fc-43ab-ad26-42ae80715706-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.550832 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-th89f\" (UniqueName: \"kubernetes.io/projected/eeb3879e-b8fc-43ab-ad26-42ae80715706-kube-api-access-th89f\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.550843 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eeb3879e-b8fc-43ab-ad26-42ae80715706-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.553914 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/585be2c9-64da-4716-9992-00e1c28380e5-kube-api-access-58f8m" (OuterVolumeSpecName: "kube-api-access-58f8m") pod "585be2c9-64da-4716-9992-00e1c28380e5" (UID: "585be2c9-64da-4716-9992-00e1c28380e5"). InnerVolumeSpecName "kube-api-access-58f8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.555447 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/585be2c9-64da-4716-9992-00e1c28380e5-logs" (OuterVolumeSpecName: "logs") pod "585be2c9-64da-4716-9992-00e1c28380e5" (UID: "585be2c9-64da-4716-9992-00e1c28380e5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.576645 4934 scope.go:117] "RemoveContainer" containerID="64f3c0c71c1d78ed7ca72c09a9c0b714710b249faec05fe75efdacefc4e67628" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.604205 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.604183468 podStartE2EDuration="2.604183468s" podCreationTimestamp="2025-10-02 10:10:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:10:28.570864098 +0000 UTC m=+1300.323505620" watchObservedRunningTime="2025-10-02 10:10:28.604183468 +0000 UTC m=+1300.356825000" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.606771 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/585be2c9-64da-4716-9992-00e1c28380e5-config-data" (OuterVolumeSpecName: "config-data") pod "585be2c9-64da-4716-9992-00e1c28380e5" (UID: "585be2c9-64da-4716-9992-00e1c28380e5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.609841 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.612552 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/585be2c9-64da-4716-9992-00e1c28380e5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "585be2c9-64da-4716-9992-00e1c28380e5" (UID: "585be2c9-64da-4716-9992-00e1c28380e5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.617456 4934 scope.go:117] "RemoveContainer" containerID="56c15d17fc08c587b0472bc5d891bd534cfcdd556ee5bd5ae3750ee8dd3c8529" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.618321 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:10:28 crc kubenswrapper[4934]: E1002 10:10:28.622169 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56c15d17fc08c587b0472bc5d891bd534cfcdd556ee5bd5ae3750ee8dd3c8529\": container with ID starting with 56c15d17fc08c587b0472bc5d891bd534cfcdd556ee5bd5ae3750ee8dd3c8529 not found: ID does not exist" containerID="56c15d17fc08c587b0472bc5d891bd534cfcdd556ee5bd5ae3750ee8dd3c8529" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.622285 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56c15d17fc08c587b0472bc5d891bd534cfcdd556ee5bd5ae3750ee8dd3c8529"} err="failed to get container status \"56c15d17fc08c587b0472bc5d891bd534cfcdd556ee5bd5ae3750ee8dd3c8529\": rpc error: code = NotFound desc = could not find container \"56c15d17fc08c587b0472bc5d891bd534cfcdd556ee5bd5ae3750ee8dd3c8529\": container with ID starting with 56c15d17fc08c587b0472bc5d891bd534cfcdd556ee5bd5ae3750ee8dd3c8529 not found: ID does not exist" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.622332 4934 scope.go:117] "RemoveContainer" containerID="64f3c0c71c1d78ed7ca72c09a9c0b714710b249faec05fe75efdacefc4e67628" Oct 02 10:10:28 crc kubenswrapper[4934]: E1002 10:10:28.622820 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64f3c0c71c1d78ed7ca72c09a9c0b714710b249faec05fe75efdacefc4e67628\": container with ID starting with 64f3c0c71c1d78ed7ca72c09a9c0b714710b249faec05fe75efdacefc4e67628 not found: ID does not exist" containerID="64f3c0c71c1d78ed7ca72c09a9c0b714710b249faec05fe75efdacefc4e67628" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.622846 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64f3c0c71c1d78ed7ca72c09a9c0b714710b249faec05fe75efdacefc4e67628"} err="failed to get container status \"64f3c0c71c1d78ed7ca72c09a9c0b714710b249faec05fe75efdacefc4e67628\": rpc error: code = NotFound desc = could not find container \"64f3c0c71c1d78ed7ca72c09a9c0b714710b249faec05fe75efdacefc4e67628\": container with ID starting with 64f3c0c71c1d78ed7ca72c09a9c0b714710b249faec05fe75efdacefc4e67628 not found: ID does not exist" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.622861 4934 scope.go:117] "RemoveContainer" containerID="c6b414c90f3b2c8ee4f300f9213c49d4f93007c4ebe815709d38d734fde7cf3b" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.624432 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/585be2c9-64da-4716-9992-00e1c28380e5-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "585be2c9-64da-4716-9992-00e1c28380e5" (UID: "585be2c9-64da-4716-9992-00e1c28380e5"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.626179 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:10:28 crc kubenswrapper[4934]: E1002 10:10:28.626539 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eeb3879e-b8fc-43ab-ad26-42ae80715706" containerName="nova-scheduler-scheduler" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.626565 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="eeb3879e-b8fc-43ab-ad26-42ae80715706" containerName="nova-scheduler-scheduler" Oct 02 10:10:28 crc kubenswrapper[4934]: E1002 10:10:28.626614 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="585be2c9-64da-4716-9992-00e1c28380e5" containerName="nova-api-log" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.626623 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="585be2c9-64da-4716-9992-00e1c28380e5" containerName="nova-api-log" Oct 02 10:10:28 crc kubenswrapper[4934]: E1002 10:10:28.626633 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="585be2c9-64da-4716-9992-00e1c28380e5" containerName="nova-api-api" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.626641 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="585be2c9-64da-4716-9992-00e1c28380e5" containerName="nova-api-api" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.626841 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="eeb3879e-b8fc-43ab-ad26-42ae80715706" containerName="nova-scheduler-scheduler" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.626865 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="585be2c9-64da-4716-9992-00e1c28380e5" containerName="nova-api-log" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.626879 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="585be2c9-64da-4716-9992-00e1c28380e5" containerName="nova-api-api" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.627477 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.630180 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.635436 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.650920 4934 scope.go:117] "RemoveContainer" containerID="c6b414c90f3b2c8ee4f300f9213c49d4f93007c4ebe815709d38d734fde7cf3b" Oct 02 10:10:28 crc kubenswrapper[4934]: E1002 10:10:28.651506 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6b414c90f3b2c8ee4f300f9213c49d4f93007c4ebe815709d38d734fde7cf3b\": container with ID starting with c6b414c90f3b2c8ee4f300f9213c49d4f93007c4ebe815709d38d734fde7cf3b not found: ID does not exist" containerID="c6b414c90f3b2c8ee4f300f9213c49d4f93007c4ebe815709d38d734fde7cf3b" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.651547 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6b414c90f3b2c8ee4f300f9213c49d4f93007c4ebe815709d38d734fde7cf3b"} err="failed to get container status \"c6b414c90f3b2c8ee4f300f9213c49d4f93007c4ebe815709d38d734fde7cf3b\": rpc error: code = NotFound desc = could not find container \"c6b414c90f3b2c8ee4f300f9213c49d4f93007c4ebe815709d38d734fde7cf3b\": container with ID starting with c6b414c90f3b2c8ee4f300f9213c49d4f93007c4ebe815709d38d734fde7cf3b not found: ID does not exist" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.652003 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfd24\" (UniqueName: \"kubernetes.io/projected/06adab43-0177-440c-aeef-9386345160c4-kube-api-access-pfd24\") pod \"nova-scheduler-0\" (UID: \"06adab43-0177-440c-aeef-9386345160c4\") " pod="openstack/nova-scheduler-0" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.652156 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06adab43-0177-440c-aeef-9386345160c4-config-data\") pod \"nova-scheduler-0\" (UID: \"06adab43-0177-440c-aeef-9386345160c4\") " pod="openstack/nova-scheduler-0" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.652224 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06adab43-0177-440c-aeef-9386345160c4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"06adab43-0177-440c-aeef-9386345160c4\") " pod="openstack/nova-scheduler-0" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.652453 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/585be2c9-64da-4716-9992-00e1c28380e5-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.652517 4934 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/585be2c9-64da-4716-9992-00e1c28380e5-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.652526 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585be2c9-64da-4716-9992-00e1c28380e5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.652536 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58f8m\" (UniqueName: \"kubernetes.io/projected/585be2c9-64da-4716-9992-00e1c28380e5-kube-api-access-58f8m\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.652546 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/585be2c9-64da-4716-9992-00e1c28380e5-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.657596 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/585be2c9-64da-4716-9992-00e1c28380e5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "585be2c9-64da-4716-9992-00e1c28380e5" (UID: "585be2c9-64da-4716-9992-00e1c28380e5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.753377 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfd24\" (UniqueName: \"kubernetes.io/projected/06adab43-0177-440c-aeef-9386345160c4-kube-api-access-pfd24\") pod \"nova-scheduler-0\" (UID: \"06adab43-0177-440c-aeef-9386345160c4\") " pod="openstack/nova-scheduler-0" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.753477 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06adab43-0177-440c-aeef-9386345160c4-config-data\") pod \"nova-scheduler-0\" (UID: \"06adab43-0177-440c-aeef-9386345160c4\") " pod="openstack/nova-scheduler-0" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.753526 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06adab43-0177-440c-aeef-9386345160c4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"06adab43-0177-440c-aeef-9386345160c4\") " pod="openstack/nova-scheduler-0" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.753598 4934 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/585be2c9-64da-4716-9992-00e1c28380e5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.756632 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06adab43-0177-440c-aeef-9386345160c4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"06adab43-0177-440c-aeef-9386345160c4\") " pod="openstack/nova-scheduler-0" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.759317 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06adab43-0177-440c-aeef-9386345160c4-config-data\") pod \"nova-scheduler-0\" (UID: \"06adab43-0177-440c-aeef-9386345160c4\") " pod="openstack/nova-scheduler-0" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.780197 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfd24\" (UniqueName: \"kubernetes.io/projected/06adab43-0177-440c-aeef-9386345160c4-kube-api-access-pfd24\") pod \"nova-scheduler-0\" (UID: \"06adab43-0177-440c-aeef-9386345160c4\") " pod="openstack/nova-scheduler-0" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.884485 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.892411 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.908847 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.910352 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.913160 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.913312 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.913339 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.924431 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="585be2c9-64da-4716-9992-00e1c28380e5" path="/var/lib/kubelet/pods/585be2c9-64da-4716-9992-00e1c28380e5/volumes" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.925483 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eeb3879e-b8fc-43ab-ad26-42ae80715706" path="/var/lib/kubelet/pods/eeb3879e-b8fc-43ab-ad26-42ae80715706/volumes" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.935616 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.945365 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.957384 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\") " pod="openstack/nova-api-0" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.957429 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2pdb\" (UniqueName: \"kubernetes.io/projected/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-kube-api-access-r2pdb\") pod \"nova-api-0\" (UID: \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\") " pod="openstack/nova-api-0" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.957486 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\") " pod="openstack/nova-api-0" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.957631 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-logs\") pod \"nova-api-0\" (UID: \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\") " pod="openstack/nova-api-0" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.957648 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-public-tls-certs\") pod \"nova-api-0\" (UID: \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\") " pod="openstack/nova-api-0" Oct 02 10:10:28 crc kubenswrapper[4934]: I1002 10:10:28.957677 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-config-data\") pod \"nova-api-0\" (UID: \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\") " pod="openstack/nova-api-0" Oct 02 10:10:29 crc kubenswrapper[4934]: I1002 10:10:29.059089 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2pdb\" (UniqueName: \"kubernetes.io/projected/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-kube-api-access-r2pdb\") pod \"nova-api-0\" (UID: \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\") " pod="openstack/nova-api-0" Oct 02 10:10:29 crc kubenswrapper[4934]: I1002 10:10:29.059152 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\") " pod="openstack/nova-api-0" Oct 02 10:10:29 crc kubenswrapper[4934]: I1002 10:10:29.059232 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-logs\") pod \"nova-api-0\" (UID: \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\") " pod="openstack/nova-api-0" Oct 02 10:10:29 crc kubenswrapper[4934]: I1002 10:10:29.059252 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-public-tls-certs\") pod \"nova-api-0\" (UID: \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\") " pod="openstack/nova-api-0" Oct 02 10:10:29 crc kubenswrapper[4934]: I1002 10:10:29.059282 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-config-data\") pod \"nova-api-0\" (UID: \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\") " pod="openstack/nova-api-0" Oct 02 10:10:29 crc kubenswrapper[4934]: I1002 10:10:29.059332 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\") " pod="openstack/nova-api-0" Oct 02 10:10:29 crc kubenswrapper[4934]: I1002 10:10:29.060178 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-logs\") pod \"nova-api-0\" (UID: \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\") " pod="openstack/nova-api-0" Oct 02 10:10:29 crc kubenswrapper[4934]: I1002 10:10:29.064398 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\") " pod="openstack/nova-api-0" Oct 02 10:10:29 crc kubenswrapper[4934]: I1002 10:10:29.064618 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-config-data\") pod \"nova-api-0\" (UID: \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\") " pod="openstack/nova-api-0" Oct 02 10:10:29 crc kubenswrapper[4934]: I1002 10:10:29.065431 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\") " pod="openstack/nova-api-0" Oct 02 10:10:29 crc kubenswrapper[4934]: I1002 10:10:29.079546 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-public-tls-certs\") pod \"nova-api-0\" (UID: \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\") " pod="openstack/nova-api-0" Oct 02 10:10:29 crc kubenswrapper[4934]: I1002 10:10:29.081148 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2pdb\" (UniqueName: \"kubernetes.io/projected/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-kube-api-access-r2pdb\") pod \"nova-api-0\" (UID: \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\") " pod="openstack/nova-api-0" Oct 02 10:10:29 crc kubenswrapper[4934]: I1002 10:10:29.232844 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:10:29 crc kubenswrapper[4934]: I1002 10:10:29.393618 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:10:29 crc kubenswrapper[4934]: W1002 10:10:29.399003 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06adab43_0177_440c_aeef_9386345160c4.slice/crio-617ae8bc456aa9d23f135d10b5fd03a0b75adf6e6fb088e0d1779fd1c7dc57ba WatchSource:0}: Error finding container 617ae8bc456aa9d23f135d10b5fd03a0b75adf6e6fb088e0d1779fd1c7dc57ba: Status 404 returned error can't find the container with id 617ae8bc456aa9d23f135d10b5fd03a0b75adf6e6fb088e0d1779fd1c7dc57ba Oct 02 10:10:29 crc kubenswrapper[4934]: I1002 10:10:29.565436 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"06adab43-0177-440c-aeef-9386345160c4","Type":"ContainerStarted","Data":"617ae8bc456aa9d23f135d10b5fd03a0b75adf6e6fb088e0d1779fd1c7dc57ba"} Oct 02 10:10:29 crc kubenswrapper[4934]: I1002 10:10:29.738262 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:10:29 crc kubenswrapper[4934]: W1002 10:10:29.745063 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4709e8f_24b0_48e6_ab94_39ec23f9ff64.slice/crio-6002a20d934ea69674bd3c416c4f994aa3e745c4fca0dd96c079749fe8cfa0ca WatchSource:0}: Error finding container 6002a20d934ea69674bd3c416c4f994aa3e745c4fca0dd96c079749fe8cfa0ca: Status 404 returned error can't find the container with id 6002a20d934ea69674bd3c416c4f994aa3e745c4fca0dd96c079749fe8cfa0ca Oct 02 10:10:30 crc kubenswrapper[4934]: I1002 10:10:30.589657 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d4709e8f-24b0-48e6-ab94-39ec23f9ff64","Type":"ContainerStarted","Data":"ad980858f1e631f4751cce1464f341435468c25a526741faf7b3a1d60a45898a"} Oct 02 10:10:30 crc kubenswrapper[4934]: I1002 10:10:30.592522 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d4709e8f-24b0-48e6-ab94-39ec23f9ff64","Type":"ContainerStarted","Data":"b19b0bf49673fda59a55fb96cbe24384993711b894ffa56c5b3e5b90372c5231"} Oct 02 10:10:30 crc kubenswrapper[4934]: I1002 10:10:30.592815 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d4709e8f-24b0-48e6-ab94-39ec23f9ff64","Type":"ContainerStarted","Data":"6002a20d934ea69674bd3c416c4f994aa3e745c4fca0dd96c079749fe8cfa0ca"} Oct 02 10:10:30 crc kubenswrapper[4934]: I1002 10:10:30.594994 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"06adab43-0177-440c-aeef-9386345160c4","Type":"ContainerStarted","Data":"b7f8b726c37d4fd8d67700b1306d05906985514bcb56092fd5f3e2779cf2dc8a"} Oct 02 10:10:30 crc kubenswrapper[4934]: I1002 10:10:30.625525 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.625468797 podStartE2EDuration="2.625468797s" podCreationTimestamp="2025-10-02 10:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:10:30.614274568 +0000 UTC m=+1302.366916100" watchObservedRunningTime="2025-10-02 10:10:30.625468797 +0000 UTC m=+1302.378110339" Oct 02 10:10:30 crc kubenswrapper[4934]: I1002 10:10:30.641404 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.641366466 podStartE2EDuration="2.641366466s" podCreationTimestamp="2025-10-02 10:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:10:30.638202489 +0000 UTC m=+1302.390844071" watchObservedRunningTime="2025-10-02 10:10:30.641366466 +0000 UTC m=+1302.394008028" Oct 02 10:10:32 crc kubenswrapper[4934]: I1002 10:10:32.184238 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 10:10:32 crc kubenswrapper[4934]: I1002 10:10:32.184623 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 10:10:33 crc kubenswrapper[4934]: I1002 10:10:33.946260 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 10:10:37 crc kubenswrapper[4934]: I1002 10:10:37.184014 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 10:10:37 crc kubenswrapper[4934]: I1002 10:10:37.184669 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 10:10:38 crc kubenswrapper[4934]: I1002 10:10:38.199891 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="04aff378-509c-49be-bf4c-2d59111fc910" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 10:10:38 crc kubenswrapper[4934]: I1002 10:10:38.199912 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="04aff378-509c-49be-bf4c-2d59111fc910" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 10:10:38 crc kubenswrapper[4934]: I1002 10:10:38.439845 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:10:38 crc kubenswrapper[4934]: I1002 10:10:38.439909 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:10:38 crc kubenswrapper[4934]: I1002 10:10:38.945971 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 10:10:38 crc kubenswrapper[4934]: I1002 10:10:38.976955 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 10:10:39 crc kubenswrapper[4934]: I1002 10:10:39.234186 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 10:10:39 crc kubenswrapper[4934]: I1002 10:10:39.234652 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 10:10:39 crc kubenswrapper[4934]: I1002 10:10:39.759281 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 10:10:40 crc kubenswrapper[4934]: I1002 10:10:40.254747 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d4709e8f-24b0-48e6-ab94-39ec23f9ff64" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.205:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 10:10:40 crc kubenswrapper[4934]: I1002 10:10:40.254829 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d4709e8f-24b0-48e6-ab94-39ec23f9ff64" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.205:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 10:10:46 crc kubenswrapper[4934]: I1002 10:10:46.757086 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 10:10:47 crc kubenswrapper[4934]: I1002 10:10:47.198891 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 10:10:47 crc kubenswrapper[4934]: I1002 10:10:47.214099 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 10:10:47 crc kubenswrapper[4934]: I1002 10:10:47.214474 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 10:10:47 crc kubenswrapper[4934]: I1002 10:10:47.823886 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 10:10:49 crc kubenswrapper[4934]: I1002 10:10:49.252183 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 10:10:49 crc kubenswrapper[4934]: I1002 10:10:49.253162 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 10:10:49 crc kubenswrapper[4934]: I1002 10:10:49.253420 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 10:10:49 crc kubenswrapper[4934]: I1002 10:10:49.261189 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 10:10:49 crc kubenswrapper[4934]: I1002 10:10:49.837289 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 10:10:49 crc kubenswrapper[4934]: I1002 10:10:49.844059 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 10:11:04 crc kubenswrapper[4934]: I1002 10:11:04.359376 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cgkdl"] Oct 02 10:11:04 crc kubenswrapper[4934]: I1002 10:11:04.370286 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cgkdl" Oct 02 10:11:04 crc kubenswrapper[4934]: I1002 10:11:04.371005 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cgkdl"] Oct 02 10:11:04 crc kubenswrapper[4934]: I1002 10:11:04.424645 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/029cff51-0959-4fdc-ba9e-44d61abbdf09-catalog-content\") pod \"community-operators-cgkdl\" (UID: \"029cff51-0959-4fdc-ba9e-44d61abbdf09\") " pod="openshift-marketplace/community-operators-cgkdl" Oct 02 10:11:04 crc kubenswrapper[4934]: I1002 10:11:04.424698 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8cw7\" (UniqueName: \"kubernetes.io/projected/029cff51-0959-4fdc-ba9e-44d61abbdf09-kube-api-access-l8cw7\") pod \"community-operators-cgkdl\" (UID: \"029cff51-0959-4fdc-ba9e-44d61abbdf09\") " pod="openshift-marketplace/community-operators-cgkdl" Oct 02 10:11:04 crc kubenswrapper[4934]: I1002 10:11:04.424751 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/029cff51-0959-4fdc-ba9e-44d61abbdf09-utilities\") pod \"community-operators-cgkdl\" (UID: \"029cff51-0959-4fdc-ba9e-44d61abbdf09\") " pod="openshift-marketplace/community-operators-cgkdl" Oct 02 10:11:04 crc kubenswrapper[4934]: I1002 10:11:04.527334 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8cw7\" (UniqueName: \"kubernetes.io/projected/029cff51-0959-4fdc-ba9e-44d61abbdf09-kube-api-access-l8cw7\") pod \"community-operators-cgkdl\" (UID: \"029cff51-0959-4fdc-ba9e-44d61abbdf09\") " pod="openshift-marketplace/community-operators-cgkdl" Oct 02 10:11:04 crc kubenswrapper[4934]: I1002 10:11:04.527368 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/029cff51-0959-4fdc-ba9e-44d61abbdf09-catalog-content\") pod \"community-operators-cgkdl\" (UID: \"029cff51-0959-4fdc-ba9e-44d61abbdf09\") " pod="openshift-marketplace/community-operators-cgkdl" Oct 02 10:11:04 crc kubenswrapper[4934]: I1002 10:11:04.527404 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/029cff51-0959-4fdc-ba9e-44d61abbdf09-utilities\") pod \"community-operators-cgkdl\" (UID: \"029cff51-0959-4fdc-ba9e-44d61abbdf09\") " pod="openshift-marketplace/community-operators-cgkdl" Oct 02 10:11:04 crc kubenswrapper[4934]: I1002 10:11:04.527918 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/029cff51-0959-4fdc-ba9e-44d61abbdf09-catalog-content\") pod \"community-operators-cgkdl\" (UID: \"029cff51-0959-4fdc-ba9e-44d61abbdf09\") " pod="openshift-marketplace/community-operators-cgkdl" Oct 02 10:11:04 crc kubenswrapper[4934]: I1002 10:11:04.527958 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/029cff51-0959-4fdc-ba9e-44d61abbdf09-utilities\") pod \"community-operators-cgkdl\" (UID: \"029cff51-0959-4fdc-ba9e-44d61abbdf09\") " pod="openshift-marketplace/community-operators-cgkdl" Oct 02 10:11:04 crc kubenswrapper[4934]: I1002 10:11:04.547349 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8cw7\" (UniqueName: \"kubernetes.io/projected/029cff51-0959-4fdc-ba9e-44d61abbdf09-kube-api-access-l8cw7\") pod \"community-operators-cgkdl\" (UID: \"029cff51-0959-4fdc-ba9e-44d61abbdf09\") " pod="openshift-marketplace/community-operators-cgkdl" Oct 02 10:11:04 crc kubenswrapper[4934]: I1002 10:11:04.744250 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cgkdl" Oct 02 10:11:05 crc kubenswrapper[4934]: I1002 10:11:05.221456 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cgkdl"] Oct 02 10:11:06 crc kubenswrapper[4934]: I1002 10:11:06.021790 4934 generic.go:334] "Generic (PLEG): container finished" podID="029cff51-0959-4fdc-ba9e-44d61abbdf09" containerID="d98b6d47a7c68c173a17fd3abe6da8ef4c4f1604d284e9691179197a36263ad9" exitCode=0 Oct 02 10:11:06 crc kubenswrapper[4934]: I1002 10:11:06.021846 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgkdl" event={"ID":"029cff51-0959-4fdc-ba9e-44d61abbdf09","Type":"ContainerDied","Data":"d98b6d47a7c68c173a17fd3abe6da8ef4c4f1604d284e9691179197a36263ad9"} Oct 02 10:11:06 crc kubenswrapper[4934]: I1002 10:11:06.021876 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgkdl" event={"ID":"029cff51-0959-4fdc-ba9e-44d61abbdf09","Type":"ContainerStarted","Data":"5c30b4bd89ca50f3b9929d6145b7904e828e61ec6c78e5bb156612d78147c723"} Oct 02 10:11:08 crc kubenswrapper[4934]: I1002 10:11:08.043558 4934 generic.go:334] "Generic (PLEG): container finished" podID="029cff51-0959-4fdc-ba9e-44d61abbdf09" containerID="04dbc5eb11c41f008a7566a44294731392b57cf65c3b3f5edddf43c9ef5d5cf2" exitCode=0 Oct 02 10:11:08 crc kubenswrapper[4934]: I1002 10:11:08.043612 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgkdl" event={"ID":"029cff51-0959-4fdc-ba9e-44d61abbdf09","Type":"ContainerDied","Data":"04dbc5eb11c41f008a7566a44294731392b57cf65c3b3f5edddf43c9ef5d5cf2"} Oct 02 10:11:08 crc kubenswrapper[4934]: I1002 10:11:08.439472 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:11:08 crc kubenswrapper[4934]: I1002 10:11:08.439804 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:11:08 crc kubenswrapper[4934]: I1002 10:11:08.439854 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 10:11:08 crc kubenswrapper[4934]: I1002 10:11:08.440691 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"577a0f8f7b9147143b952a1ec195b9368082248053d4e85cde27a2757e47ab8c"} pod="openshift-machine-config-operator/machine-config-daemon-djh5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 10:11:08 crc kubenswrapper[4934]: I1002 10:11:08.440768 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" containerID="cri-o://577a0f8f7b9147143b952a1ec195b9368082248053d4e85cde27a2757e47ab8c" gracePeriod=600 Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.076056 4934 generic.go:334] "Generic (PLEG): container finished" podID="71db06ef-05b0-4f58-b251-b27117a8500a" containerID="577a0f8f7b9147143b952a1ec195b9368082248053d4e85cde27a2757e47ab8c" exitCode=0 Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.076326 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerDied","Data":"577a0f8f7b9147143b952a1ec195b9368082248053d4e85cde27a2757e47ab8c"} Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.076350 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerStarted","Data":"f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1"} Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.076367 4934 scope.go:117] "RemoveContainer" containerID="7b2842dc9e6894e9abab64461b922632ce5e3be7a7134bbbd340310d8f4b1d25" Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.110030 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgkdl" event={"ID":"029cff51-0959-4fdc-ba9e-44d61abbdf09","Type":"ContainerStarted","Data":"a4c2d36a689c80cbdf56b5ef00161c7b75929cbc632517514af288d8cc84f00a"} Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.285632 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cgkdl" podStartSLOduration=2.515165578 podStartE2EDuration="5.285616039s" podCreationTimestamp="2025-10-02 10:11:04 +0000 UTC" firstStartedPulling="2025-10-02 10:11:06.024205094 +0000 UTC m=+1337.776846626" lastFinishedPulling="2025-10-02 10:11:08.794655555 +0000 UTC m=+1340.547297087" observedRunningTime="2025-10-02 10:11:09.191041058 +0000 UTC m=+1340.943682580" watchObservedRunningTime="2025-10-02 10:11:09.285616039 +0000 UTC m=+1341.038257561" Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.299287 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.299573 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="9c99e264-05b3-4cd7-91ae-cabb6d10466a" containerName="openstackclient" containerID="cri-o://7e1c034adb12673de1bc4203a0717c7690b9b6165dd60bea908b057b397ba31e" gracePeriod=2 Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.309069 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.366489 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 10:11:09 crc kubenswrapper[4934]: E1002 10:11:09.538230 4934 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 02 10:11:09 crc kubenswrapper[4934]: E1002 10:11:09.538547 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d2b6e083-a523-48e2-8157-088036174e12-config-data podName:d2b6e083-a523-48e2-8157-088036174e12 nodeName:}" failed. No retries permitted until 2025-10-02 10:11:10.038529971 +0000 UTC m=+1341.791171493 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/d2b6e083-a523-48e2-8157-088036174e12-config-data") pod "rabbitmq-cell1-server-0" (UID: "d2b6e083-a523-48e2-8157-088036174e12") : configmap "rabbitmq-cell1-config-data" not found Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.575537 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.575779 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="05c8faa0-6986-415d-abc3-4d35291375f2" containerName="ovn-northd" containerID="cri-o://6ab70322bf5ef2e1d2a4fd23bb06ac7d14bc671fa582c8397ceb83e14554def5" gracePeriod=30 Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.576168 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="05c8faa0-6986-415d-abc3-4d35291375f2" containerName="openstack-network-exporter" containerID="cri-o://c0691673552ec68e68f4be61b4e9826705cb137abb4ac7f57faa0ea2d303510d" gracePeriod=30 Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.732052 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.732647 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="9d816754-2d4c-40e6-acee-218947ab9e19" containerName="openstack-network-exporter" containerID="cri-o://66355c51e4027fd1c4a3713d064cb84130067b07985283b47ad3d5d705b2d508" gracePeriod=300 Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.754813 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance82a2-account-delete-sv9vf"] Oct 02 10:11:09 crc kubenswrapper[4934]: E1002 10:11:09.755239 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c99e264-05b3-4cd7-91ae-cabb6d10466a" containerName="openstackclient" Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.755257 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c99e264-05b3-4cd7-91ae-cabb6d10466a" containerName="openstackclient" Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.755458 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c99e264-05b3-4cd7-91ae-cabb6d10466a" containerName="openstackclient" Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.756084 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance82a2-account-delete-sv9vf" Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.777118 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance82a2-account-delete-sv9vf"] Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.824725 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder36fe-account-delete-n2bk5"] Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.842659 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder36fe-account-delete-n2bk5" Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.852972 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tqdj\" (UniqueName: \"kubernetes.io/projected/8679bb63-5c56-42a3-9b2c-42303e3afef6-kube-api-access-8tqdj\") pod \"glance82a2-account-delete-sv9vf\" (UID: \"8679bb63-5c56-42a3-9b2c-42303e3afef6\") " pod="openstack/glance82a2-account-delete-sv9vf" Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.877267 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder36fe-account-delete-n2bk5"] Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.893979 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="9d816754-2d4c-40e6-acee-218947ab9e19" containerName="ovsdbserver-nb" containerID="cri-o://be30a7274c35eb6dd2db94c2805716e312b3406c00793273d691971991038504" gracePeriod=300 Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.941241 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-l476l"] Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.941480 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-l476l" podUID="e14738be-f83f-47ce-b0d9-ceb23d504b46" containerName="openstack-network-exporter" containerID="cri-o://336b71c58bd0e5981fd64408bac6c360461733df6dbdbba45a423f97c6077a34" gracePeriod=30 Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.956374 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement5c1f-account-delete-2jjrd"] Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.958357 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement5c1f-account-delete-2jjrd" Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.961898 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement5c1f-account-delete-2jjrd"] Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.978538 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-gwpkm"] Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.983535 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tqdj\" (UniqueName: \"kubernetes.io/projected/8679bb63-5c56-42a3-9b2c-42303e3afef6-kube-api-access-8tqdj\") pod \"glance82a2-account-delete-sv9vf\" (UID: \"8679bb63-5c56-42a3-9b2c-42303e3afef6\") " pod="openstack/glance82a2-account-delete-sv9vf" Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.983611 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hj7sq\" (UniqueName: \"kubernetes.io/projected/bf007c3a-2ba5-4d94-a789-a5413787d6d8-kube-api-access-hj7sq\") pod \"cinder36fe-account-delete-n2bk5\" (UID: \"bf007c3a-2ba5-4d94-a789-a5413787d6d8\") " pod="openstack/cinder36fe-account-delete-n2bk5" Oct 02 10:11:09 crc kubenswrapper[4934]: I1002 10:11:09.983916 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-65wtc"] Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.053246 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-rqm5f"] Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.066561 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-rqm5f"] Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.081427 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tqdj\" (UniqueName: \"kubernetes.io/projected/8679bb63-5c56-42a3-9b2c-42303e3afef6-kube-api-access-8tqdj\") pod \"glance82a2-account-delete-sv9vf\" (UID: \"8679bb63-5c56-42a3-9b2c-42303e3afef6\") " pod="openstack/glance82a2-account-delete-sv9vf" Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.090973 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hj7sq\" (UniqueName: \"kubernetes.io/projected/bf007c3a-2ba5-4d94-a789-a5413787d6d8-kube-api-access-hj7sq\") pod \"cinder36fe-account-delete-n2bk5\" (UID: \"bf007c3a-2ba5-4d94-a789-a5413787d6d8\") " pod="openstack/cinder36fe-account-delete-n2bk5" Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.091999 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44tbd\" (UniqueName: \"kubernetes.io/projected/5ceb70f8-5cc3-4e91-9bc1-6fd0b7575efc-kube-api-access-44tbd\") pod \"placement5c1f-account-delete-2jjrd\" (UID: \"5ceb70f8-5cc3-4e91-9bc1-6fd0b7575efc\") " pod="openstack/placement5c1f-account-delete-2jjrd" Oct 02 10:11:10 crc kubenswrapper[4934]: E1002 10:11:10.092098 4934 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 02 10:11:10 crc kubenswrapper[4934]: E1002 10:11:10.092145 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d2b6e083-a523-48e2-8157-088036174e12-config-data podName:d2b6e083-a523-48e2-8157-088036174e12 nodeName:}" failed. No retries permitted until 2025-10-02 10:11:11.092129934 +0000 UTC m=+1342.844771456 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/d2b6e083-a523-48e2-8157-088036174e12-config-data") pod "rabbitmq-cell1-server-0" (UID: "d2b6e083-a523-48e2-8157-088036174e12") : configmap "rabbitmq-cell1-config-data" not found Oct 02 10:11:10 crc kubenswrapper[4934]: E1002 10:11:10.093028 4934 projected.go:263] Couldn't get secret openstack/swift-conf: secret "swift-conf" not found Oct 02 10:11:10 crc kubenswrapper[4934]: E1002 10:11:10.093046 4934 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-proxy-bfc5bc4b9-crmsr: secret "swift-conf" not found Oct 02 10:11:10 crc kubenswrapper[4934]: E1002 10:11:10.093070 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a30daf4c-06f2-48f5-b57f-f6704fae13a9-etc-swift podName:a30daf4c-06f2-48f5-b57f-f6704fae13a9 nodeName:}" failed. No retries permitted until 2025-10-02 10:11:10.593061069 +0000 UTC m=+1342.345702591 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a30daf4c-06f2-48f5-b57f-f6704fae13a9-etc-swift") pod "swift-proxy-bfc5bc4b9-crmsr" (UID: "a30daf4c-06f2-48f5-b57f-f6704fae13a9") : secret "swift-conf" not found Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.129243 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hj7sq\" (UniqueName: \"kubernetes.io/projected/bf007c3a-2ba5-4d94-a789-a5413787d6d8-kube-api-access-hj7sq\") pod \"cinder36fe-account-delete-n2bk5\" (UID: \"bf007c3a-2ba5-4d94-a789-a5413787d6d8\") " pod="openstack/cinder36fe-account-delete-n2bk5" Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.143158 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance82a2-account-delete-sv9vf" Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.148758 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.187932 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_9d816754-2d4c-40e6-acee-218947ab9e19/ovsdbserver-nb/0.log" Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.187977 4934 generic.go:334] "Generic (PLEG): container finished" podID="9d816754-2d4c-40e6-acee-218947ab9e19" containerID="be30a7274c35eb6dd2db94c2805716e312b3406c00793273d691971991038504" exitCode=143 Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.188032 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"9d816754-2d4c-40e6-acee-218947ab9e19","Type":"ContainerDied","Data":"be30a7274c35eb6dd2db94c2805716e312b3406c00793273d691971991038504"} Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.188640 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-jwpkw"] Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.194108 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44tbd\" (UniqueName: \"kubernetes.io/projected/5ceb70f8-5cc3-4e91-9bc1-6fd0b7575efc-kube-api-access-44tbd\") pod \"placement5c1f-account-delete-2jjrd\" (UID: \"5ceb70f8-5cc3-4e91-9bc1-6fd0b7575efc\") " pod="openstack/placement5c1f-account-delete-2jjrd" Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.205892 4934 generic.go:334] "Generic (PLEG): container finished" podID="05c8faa0-6986-415d-abc3-4d35291375f2" containerID="c0691673552ec68e68f4be61b4e9826705cb137abb4ac7f57faa0ea2d303510d" exitCode=2 Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.205997 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"05c8faa0-6986-415d-abc3-4d35291375f2","Type":"ContainerDied","Data":"c0691673552ec68e68f4be61b4e9826705cb137abb4ac7f57faa0ea2d303510d"} Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.215555 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44tbd\" (UniqueName: \"kubernetes.io/projected/5ceb70f8-5cc3-4e91-9bc1-6fd0b7575efc-kube-api-access-44tbd\") pod \"placement5c1f-account-delete-2jjrd\" (UID: \"5ceb70f8-5cc3-4e91-9bc1-6fd0b7575efc\") " pod="openstack/placement5c1f-account-delete-2jjrd" Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.252300 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-jwpkw"] Oct 02 10:11:10 crc kubenswrapper[4934]: E1002 10:11:10.296685 4934 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 02 10:11:10 crc kubenswrapper[4934]: E1002 10:11:10.297735 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/97560469-c459-48f4-901e-54eaded32537-config-data podName:97560469-c459-48f4-901e-54eaded32537 nodeName:}" failed. No retries permitted until 2025-10-02 10:11:10.797719419 +0000 UTC m=+1342.550360941 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/97560469-c459-48f4-901e-54eaded32537-config-data") pod "rabbitmq-server-0" (UID: "97560469-c459-48f4-901e-54eaded32537") : configmap "rabbitmq-config-data" not found Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.332202 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement5c1f-account-delete-2jjrd" Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.358178 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder36fe-account-delete-n2bk5" Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.422193 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapi26a6-account-delete-ng2mb"] Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.423435 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi26a6-account-delete-ng2mb" Oct 02 10:11:10 crc kubenswrapper[4934]: E1002 10:11:10.434106 4934 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err="command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: " execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-gwpkm" message=< Oct 02 10:11:10 crc kubenswrapper[4934]: Exiting ovn-controller (1) [ OK ] Oct 02 10:11:10 crc kubenswrapper[4934]: > Oct 02 10:11:10 crc kubenswrapper[4934]: E1002 10:11:10.434144 4934 kuberuntime_container.go:691] "PreStop hook failed" err="command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: " pod="openstack/ovn-controller-gwpkm" podUID="7214291c-3c8e-4173-a6fd-ef16170258e2" containerName="ovn-controller" containerID="cri-o://5ab65d1f2a5fdc33711875776226a728389485f48a057c73387e2db56049cbbc" Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.434175 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-gwpkm" podUID="7214291c-3c8e-4173-a6fd-ef16170258e2" containerName="ovn-controller" containerID="cri-o://5ab65d1f2a5fdc33711875776226a728389485f48a057c73387e2db56049cbbc" gracePeriod=30 Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.477457 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-kb5lq"] Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.498820 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-kb5lq"] Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.503023 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlc5h\" (UniqueName: \"kubernetes.io/projected/a7f7a4d6-c246-4c88-85e1-573fd07ec33f-kube-api-access-jlc5h\") pod \"novaapi26a6-account-delete-ng2mb\" (UID: \"a7f7a4d6-c246-4c88-85e1-573fd07ec33f\") " pod="openstack/novaapi26a6-account-delete-ng2mb" Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.531465 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapi26a6-account-delete-ng2mb"] Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.552688 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-klshd"] Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.558366 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-klshd"] Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.604788 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlc5h\" (UniqueName: \"kubernetes.io/projected/a7f7a4d6-c246-4c88-85e1-573fd07ec33f-kube-api-access-jlc5h\") pod \"novaapi26a6-account-delete-ng2mb\" (UID: \"a7f7a4d6-c246-4c88-85e1-573fd07ec33f\") " pod="openstack/novaapi26a6-account-delete-ng2mb" Oct 02 10:11:10 crc kubenswrapper[4934]: E1002 10:11:10.605132 4934 projected.go:263] Couldn't get secret openstack/swift-proxy-config-data: secret "swift-proxy-config-data" not found Oct 02 10:11:10 crc kubenswrapper[4934]: E1002 10:11:10.605153 4934 projected.go:263] Couldn't get secret openstack/swift-conf: secret "swift-conf" not found Oct 02 10:11:10 crc kubenswrapper[4934]: E1002 10:11:10.605164 4934 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 10:11:10 crc kubenswrapper[4934]: E1002 10:11:10.605179 4934 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-proxy-bfc5bc4b9-crmsr: [secret "swift-proxy-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Oct 02 10:11:10 crc kubenswrapper[4934]: E1002 10:11:10.605228 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a30daf4c-06f2-48f5-b57f-f6704fae13a9-etc-swift podName:a30daf4c-06f2-48f5-b57f-f6704fae13a9 nodeName:}" failed. No retries permitted until 2025-10-02 10:11:11.605211198 +0000 UTC m=+1343.357852720 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a30daf4c-06f2-48f5-b57f-f6704fae13a9-etc-swift") pod "swift-proxy-bfc5bc4b9-crmsr" (UID: "a30daf4c-06f2-48f5-b57f-f6704fae13a9") : [secret "swift-proxy-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.662161 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlc5h\" (UniqueName: \"kubernetes.io/projected/a7f7a4d6-c246-4c88-85e1-573fd07ec33f-kube-api-access-jlc5h\") pod \"novaapi26a6-account-delete-ng2mb\" (UID: \"a7f7a4d6-c246-4c88-85e1-573fd07ec33f\") " pod="openstack/novaapi26a6-account-delete-ng2mb" Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.759858 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-9z9t6"] Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.780400 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi26a6-account-delete-ng2mb" Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.784648 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-9z9t6"] Oct 02 10:11:10 crc kubenswrapper[4934]: E1002 10:11:10.823305 4934 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 02 10:11:10 crc kubenswrapper[4934]: E1002 10:11:10.823373 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/97560469-c459-48f4-901e-54eaded32537-config-data podName:97560469-c459-48f4-901e-54eaded32537 nodeName:}" failed. No retries permitted until 2025-10-02 10:11:11.82335948 +0000 UTC m=+1343.576001002 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/97560469-c459-48f4-901e-54eaded32537-config-data") pod "rabbitmq-server-0" (UID: "97560469-c459-48f4-901e-54eaded32537") : configmap "rabbitmq-config-data" not found Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.912401 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.913367 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="6061002c-7242-4d02-b868-6838b39f30fa" containerName="openstack-network-exporter" containerID="cri-o://fd9f74a1d204e8b395d9002c1eb07158525244b64cff40fb8c86051204dcedaa" gracePeriod=300 Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.977490 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="087ad031-cdf0-45e4-b1cf-ae777e8cc5fb" path="/var/lib/kubelet/pods/087ad031-cdf0-45e4-b1cf-ae777e8cc5fb/volumes" Oct 02 10:11:10 crc kubenswrapper[4934]: I1002 10:11:10.978275 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="157aa366-544e-42ba-bd81-1649575c976f" path="/var/lib/kubelet/pods/157aa366-544e-42ba-bd81-1649575c976f/volumes" Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.006135 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="379bc476-7f65-408d-929a-8badb1e4dfdc" path="/var/lib/kubelet/pods/379bc476-7f65-408d-929a-8badb1e4dfdc/volumes" Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.006731 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c03e58b-5a31-4295-ba05-d219daff78b1" path="/var/lib/kubelet/pods/3c03e58b-5a31-4295-ba05-d219daff78b1/volumes" Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.007234 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6427286d-a1ca-43ca-8ecb-bbdda45e360e" path="/var/lib/kubelet/pods/6427286d-a1ca-43ca-8ecb-bbdda45e360e/volumes" Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.007944 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85b4cf56ff-zr6h2"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.008150 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" podUID="20b39d07-1559-4d11-b9ff-6ef97142a58a" containerName="dnsmasq-dns" containerID="cri-o://aac633fecdf45004f95036aeac08948cfcfe75a04ef1942aef60308cc2356517" gracePeriod=10 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.014154 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-62a6-account-create-nthhn"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.079881 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-62a6-account-create-nthhn"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.089765 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-qm5fm"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.100805 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-qm5fm"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.117880 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.118120 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="94a95a16-6bdd-4c08-8340-13fd85455b7a" containerName="cinder-scheduler" containerID="cri-o://9c952faa06a3569c88d733912510a64a2634e55c3359c72da98dba68500b8acf" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.118498 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="94a95a16-6bdd-4c08-8340-13fd85455b7a" containerName="probe" containerID="cri-o://7f805c812aa2a019f6996c4840760eec127d8f206fe82dbfef04db5d8dadd02c" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: E1002 10:11:11.139852 4934 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 02 10:11:11 crc kubenswrapper[4934]: E1002 10:11:11.140139 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d2b6e083-a523-48e2-8157-088036174e12-config-data podName:d2b6e083-a523-48e2-8157-088036174e12 nodeName:}" failed. No retries permitted until 2025-10-02 10:11:13.140124945 +0000 UTC m=+1344.892766467 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/d2b6e083-a523-48e2-8157-088036174e12-config-data") pod "rabbitmq-cell1-server-0" (UID: "d2b6e083-a523-48e2-8157-088036174e12") : configmap "rabbitmq-cell1-config-data" not found Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.172613 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-zkqqc"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.193832 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-zkqqc"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.227105 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.227412 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="84d2000a-d99f-4320-b178-874940ab7e9d" containerName="glance-log" containerID="cri-o://9799cb2e3daec429b427c098af4fc69103ad295acfa71492709948a69984306d" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.227602 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="84d2000a-d99f-4320-b178-874940ab7e9d" containerName="glance-httpd" containerID="cri-o://391ab06a0bf5924bac18ac518442b3a5539dd7de7e5001512639e7be11a54744" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.251975 4934 generic.go:334] "Generic (PLEG): container finished" podID="6061002c-7242-4d02-b868-6838b39f30fa" containerID="fd9f74a1d204e8b395d9002c1eb07158525244b64cff40fb8c86051204dcedaa" exitCode=2 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.252053 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6061002c-7242-4d02-b868-6838b39f30fa","Type":"ContainerDied","Data":"fd9f74a1d204e8b395d9002c1eb07158525244b64cff40fb8c86051204dcedaa"} Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.263627 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-ldwrv"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.264823 4934 generic.go:334] "Generic (PLEG): container finished" podID="7214291c-3c8e-4173-a6fd-ef16170258e2" containerID="5ab65d1f2a5fdc33711875776226a728389485f48a057c73387e2db56049cbbc" exitCode=0 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.264882 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gwpkm" event={"ID":"7214291c-3c8e-4173-a6fd-ef16170258e2","Type":"ContainerDied","Data":"5ab65d1f2a5fdc33711875776226a728389485f48a057c73387e2db56049cbbc"} Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.278707 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-l476l_e14738be-f83f-47ce-b0d9-ceb23d504b46/openstack-network-exporter/0.log" Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.278747 4934 generic.go:334] "Generic (PLEG): container finished" podID="e14738be-f83f-47ce-b0d9-ceb23d504b46" containerID="336b71c58bd0e5981fd64408bac6c360461733df6dbdbba45a423f97c6077a34" exitCode=2 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.278824 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-l476l" event={"ID":"e14738be-f83f-47ce-b0d9-ceb23d504b46","Type":"ContainerDied","Data":"336b71c58bd0e5981fd64408bac6c360461733df6dbdbba45a423f97c6077a34"} Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.280841 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-ldwrv"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.300106 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_9d816754-2d4c-40e6-acee-218947ab9e19/ovsdbserver-nb/0.log" Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.300167 4934 generic.go:334] "Generic (PLEG): container finished" podID="9d816754-2d4c-40e6-acee-218947ab9e19" containerID="66355c51e4027fd1c4a3713d064cb84130067b07985283b47ad3d5d705b2d508" exitCode=2 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.300200 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"9d816754-2d4c-40e6-acee-218947ab9e19","Type":"ContainerDied","Data":"66355c51e4027fd1c4a3713d064cb84130067b07985283b47ad3d5d705b2d508"} Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.302442 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.307869 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="cd41dc17-6659-4b43-9758-f67d831993ba" containerName="cinder-api-log" containerID="cri-o://ddc003d07d24a576f93d547d9f6f180b883faaf22fef13259bf076d508867ebe" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.308272 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="cd41dc17-6659-4b43-9758-f67d831993ba" containerName="cinder-api" containerID="cri-o://86f78dcebecc392bfd4c80af3143e2df0ad38fe341e51b1a9c1ed70cc6f77bb2" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.337033 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-n6t65"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.345272 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="6061002c-7242-4d02-b868-6838b39f30fa" containerName="ovsdbserver-sb" containerID="cri-o://abc6a88d34ca1719e3b11b44524da713532a93344a46bb1626798816c0caa7a2" gracePeriod=300 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.348943 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-649ccd9666-dh5rs"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.349269 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-649ccd9666-dh5rs" podUID="ecc76715-8b05-4529-bd7b-289d7f32eff5" containerName="placement-log" containerID="cri-o://4d19c2050c253cd17d4e337c977a53e89370af2b71d1d6898f8b9f9d88a2f2b2" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.349645 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-649ccd9666-dh5rs" podUID="ecc76715-8b05-4529-bd7b-289d7f32eff5" containerName="placement-api" containerID="cri-o://8710e544f71f9ca7366446c35d7eb3080fa07ae1e0f59d3a870062d5e14b9eb6" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.365588 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-n6t65"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.381416 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.386807 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9165cca6-da95-45ae-87c3-b2829756db3b" containerName="glance-log" containerID="cri-o://9dde5aaa3e65635d53137792cb4e654aa61c584c4bc197377b96acd4ff67091e" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.389184 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9165cca6-da95-45ae-87c3-b2829756db3b" containerName="glance-httpd" containerID="cri-o://e38759eb2a2651b29d3a6703a0da0365f7f7843aad2b4f894f68ed24320e6d21" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.403015 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.403660 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="account-server" containerID="cri-o://478b442a459773165d46dac69a02032f32c9b37d6398eff28160ced5dd0ee4b4" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.403782 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="swift-recon-cron" containerID="cri-o://1c65da253bd0c434afbafb0cd05ae9a507db27a1ff5ce1626446a181ca57c304" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.403844 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="rsync" containerID="cri-o://9373e5fbc3b166030467efc3d48ed6860fef2654fa4d21f813440b32765b3063" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.403882 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="object-expirer" containerID="cri-o://2d1cebd6bce0741d006ac48e38db4c872771c02036472a9fa71980ddfb72cef5" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.403911 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="object-updater" containerID="cri-o://6b7e46bf6361363ea3ce0effc8f79e7c186a4d64e12c8053c444d0f89313722e" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.403942 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="object-auditor" containerID="cri-o://169e61c78b12ac121ef1306849a0fe5e05cc7125ba6f6a238a652aac3445bad8" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.403968 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="object-replicator" containerID="cri-o://d6fc4c846f604ef5014de443285ee7dab569b41691babd5a6981585d8007df21" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.404008 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="object-server" containerID="cri-o://1e2427a94989f9a77ca59046a739fc7849881116c29e79c766ddc52debf75ec3" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.404037 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="container-updater" containerID="cri-o://883824276a85ba5f915880ccf065542294a5a3959ead2ee9e02e556f4e13a641" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.404080 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="container-auditor" containerID="cri-o://022023c76cc2ba5ec65353c6ec2387833901f691565a07dc55b9a6a09bf9871c" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.404110 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="container-replicator" containerID="cri-o://da203f48b1e4a0c44c371939bd0a1b6ee7159680a2a3fac865d9476c3b7b9c05" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.404137 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="container-server" containerID="cri-o://c0f5815d02eff4beefc969cb8468e9b85a60d871f99cc29f33aa27f2cd50039f" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.404168 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="account-reaper" containerID="cri-o://65c0b15d452f5559c24817138f1d42f00b5d08d34eff5f6a036df1ded5c70ccb" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.404196 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="account-auditor" containerID="cri-o://5c785fb317467101f9ff407243fb85596511c5e525db3832b38df04ed00fec73" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.404224 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="account-replicator" containerID="cri-o://283ced1ad0dcd3181290d000eab3ac44a3033b8e32fa0ca0cd8ef03d57a7deee" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: E1002 10:11:11.501435 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5ab65d1f2a5fdc33711875776226a728389485f48a057c73387e2db56049cbbc is running failed: container process not found" containerID="5ab65d1f2a5fdc33711875776226a728389485f48a057c73387e2db56049cbbc" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.501507 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-57456ffd97-rqc8b"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.518048 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-57456ffd97-rqc8b" podUID="6d79c694-2253-419e-b63b-4a38884ac57e" containerName="neutron-api" containerID="cri-o://a68b3c26edd8aa7fbcd688303c8a83effaa456e62c9b48088363c448ed90aef6" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.521727 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-57456ffd97-rqc8b" podUID="6d79c694-2253-419e-b63b-4a38884ac57e" containerName="neutron-httpd" containerID="cri-o://959724f7a775ac28fa7ae269f3be79b301f1332c985d6f8a80a47def488f67f5" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: E1002 10:11:11.558338 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5ab65d1f2a5fdc33711875776226a728389485f48a057c73387e2db56049cbbc is running failed: container process not found" containerID="5ab65d1f2a5fdc33711875776226a728389485f48a057c73387e2db56049cbbc" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Oct 02 10:11:11 crc kubenswrapper[4934]: E1002 10:11:11.559293 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5ab65d1f2a5fdc33711875776226a728389485f48a057c73387e2db56049cbbc is running failed: container process not found" containerID="5ab65d1f2a5fdc33711875776226a728389485f48a057c73387e2db56049cbbc" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Oct 02 10:11:11 crc kubenswrapper[4934]: E1002 10:11:11.559327 4934 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5ab65d1f2a5fdc33711875776226a728389485f48a057c73387e2db56049cbbc is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-gwpkm" podUID="7214291c-3c8e-4173-a6fd-ef16170258e2" containerName="ovn-controller" Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.596695 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance82a2-account-delete-sv9vf"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.660551 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-82a2-account-create-nd4xr"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.664178 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-82a2-account-create-nd4xr"] Oct 02 10:11:11 crc kubenswrapper[4934]: E1002 10:11:11.696764 4934 projected.go:263] Couldn't get secret openstack/swift-proxy-config-data: secret "swift-proxy-config-data" not found Oct 02 10:11:11 crc kubenswrapper[4934]: E1002 10:11:11.697422 4934 projected.go:263] Couldn't get secret openstack/swift-conf: secret "swift-conf" not found Oct 02 10:11:11 crc kubenswrapper[4934]: E1002 10:11:11.697514 4934 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 10:11:11 crc kubenswrapper[4934]: E1002 10:11:11.701298 4934 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-proxy-bfc5bc4b9-crmsr: [secret "swift-proxy-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Oct 02 10:11:11 crc kubenswrapper[4934]: E1002 10:11:11.701440 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a30daf4c-06f2-48f5-b57f-f6704fae13a9-etc-swift podName:a30daf4c-06f2-48f5-b57f-f6704fae13a9 nodeName:}" failed. No retries permitted until 2025-10-02 10:11:13.701419039 +0000 UTC m=+1345.454060561 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a30daf4c-06f2-48f5-b57f-f6704fae13a9-etc-swift") pod "swift-proxy-bfc5bc4b9-crmsr" (UID: "a30daf4c-06f2-48f5-b57f-f6704fae13a9") : [secret "swift-proxy-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.709328 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-sj7h4"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.724759 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-sj7h4"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.772396 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-36fe-account-create-8rqjp"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.782313 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-cwxjc"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.797227 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder36fe-account-delete-n2bk5"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.797295 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-cwxjc"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.807960 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-36fe-account-create-8rqjp"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.817764 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-70e2-account-create-cgqd8"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.855811 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-cw4qq"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.890750 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-70e2-account-create-cgqd8"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.899426 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-5c1f-account-create-z7zpn"] Oct 02 10:11:11 crc kubenswrapper[4934]: E1002 10:11:11.905075 4934 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 02 10:11:11 crc kubenswrapper[4934]: E1002 10:11:11.905243 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/97560469-c459-48f4-901e-54eaded32537-config-data podName:97560469-c459-48f4-901e-54eaded32537 nodeName:}" failed. No retries permitted until 2025-10-02 10:11:13.905223516 +0000 UTC m=+1345.657865038 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/97560469-c459-48f4-901e-54eaded32537-config-data") pod "rabbitmq-server-0" (UID: "97560469-c459-48f4-901e-54eaded32537") : configmap "rabbitmq-config-data" not found Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.913396 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-5c1f-account-create-z7zpn"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.935050 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-cw4qq"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.946203 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-6bf48f9d67-g95nd"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.946429 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-6bf48f9d67-g95nd" podUID="547c6c6d-0f01-4feb-acc6-2e3045407b64" containerName="barbican-worker-log" containerID="cri-o://ca5c961136b99e80a8ae0bb9c203d870c7355071cc4a476bdb7c41606bdef8b7" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.946967 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-6bf48f9d67-g95nd" podUID="547c6c6d-0f01-4feb-acc6-2e3045407b64" containerName="barbican-worker" containerID="cri-o://c52d38070f0e71890f03b32d417b88c6062d9efcaddea407689d3ce9f0ed9989" gracePeriod=30 Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.967098 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-74sr7"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.974132 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement5c1f-account-delete-2jjrd"] Oct 02 10:11:11 crc kubenswrapper[4934]: I1002 10:11:11.985516 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-74sr7"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.001426 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.011226 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-8648b56f4-5f8xk"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.011444 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-8648b56f4-5f8xk" podUID="350928d4-5a0b-4331-b1c2-220517d313a5" containerName="barbican-api-log" containerID="cri-o://9245f97a19f447f9859a60a1c0d5f3f95ccff037fb5235ac35a03f447f11628f" gracePeriod=30 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.012097 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-8648b56f4-5f8xk" podUID="350928d4-5a0b-4331-b1c2-220517d313a5" containerName="barbican-api" containerID="cri-o://41b55b90784f0d27b65abc25b560172adc6a97c8f0800218d04b4ab6f18fc610" gracePeriod=30 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.025700 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-b97fc8746-vlx9z"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.025907 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-b97fc8746-vlx9z" podUID="d15b0450-e100-4ee7-ad62-88601eedc4f2" containerName="barbican-keystone-listener-log" containerID="cri-o://c83811ac76f059e3b264882bb3a6859bb99aa9ea06694e42d9967733e316ac15" gracePeriod=30 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.026262 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-b97fc8746-vlx9z" podUID="d15b0450-e100-4ee7-ad62-88601eedc4f2" containerName="barbican-keystone-listener" containerID="cri-o://2e7e1ea7602c5dbc0c4396593163b8efff5196b8cba864f827def060e37814e7" gracePeriod=30 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.044345 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.044411 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-h84gj"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.071025 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-h84gj"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.088403 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-9b8b-account-create-v55kd"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.098862 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-9b8b-account-create-v55kd"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.115121 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="d2b6e083-a523-48e2-8157-088036174e12" containerName="rabbitmq" containerID="cri-o://b2d1f1b04e17b195fbd97880fd7d7c42841c576a17071e6ee86a7e804f16c194" gracePeriod=604800 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.124813 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.125080 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d4709e8f-24b0-48e6-ab94-39ec23f9ff64" containerName="nova-api-log" containerID="cri-o://b19b0bf49673fda59a55fb96cbe24384993711b894ffa56c5b3e5b90372c5231" gracePeriod=30 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.125395 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d4709e8f-24b0-48e6-ab94-39ec23f9ff64" containerName="nova-api-api" containerID="cri-o://ad980858f1e631f4751cce1464f341435468c25a526741faf7b3a1d60a45898a" gracePeriod=30 Oct 02 10:11:12 crc kubenswrapper[4934]: E1002 10:11:12.143776 4934 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9165cca6_da95_45ae_87c3_b2829756db3b.slice/crio-9dde5aaa3e65635d53137792cb4e654aa61c584c4bc197377b96acd4ff67091e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6061002c_7242_4d02_b868_6838b39f30fa.slice/crio-abc6a88d34ca1719e3b11b44524da713532a93344a46bb1626798816c0caa7a2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6061002c_7242_4d02_b868_6838b39f30fa.slice/crio-conmon-abc6a88d34ca1719e3b11b44524da713532a93344a46bb1626798816c0caa7a2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20b39d07_1559_4d11_b9ff_6ef97142a58a.slice/crio-conmon-aac633fecdf45004f95036aeac08948cfcfe75a04ef1942aef60308cc2356517.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84d2000a_d99f_4320_b178_874940ab7e9d.slice/crio-9799cb2e3daec429b427c098af4fc69103ad295acfa71492709948a69984306d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9165cca6_da95_45ae_87c3_b2829756db3b.slice/crio-conmon-9dde5aaa3e65635d53137792cb4e654aa61c584c4bc197377b96acd4ff67091e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cebd152_5aaf_4543_be11_7ba327cfdf0e.slice/crio-6b7e46bf6361363ea3ce0effc8f79e7c186a4d64e12c8053c444d0f89313722e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cebd152_5aaf_4543_be11_7ba327cfdf0e.slice/crio-conmon-6b7e46bf6361363ea3ce0effc8f79e7c186a4d64e12c8053c444d0f89313722e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cebd152_5aaf_4543_be11_7ba327cfdf0e.slice/crio-conmon-65c0b15d452f5559c24817138f1d42f00b5d08d34eff5f6a036df1ded5c70ccb.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cebd152_5aaf_4543_be11_7ba327cfdf0e.slice/crio-conmon-283ced1ad0dcd3181290d000eab3ac44a3033b8e32fa0ca0cd8ef03d57a7deee.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cebd152_5aaf_4543_be11_7ba327cfdf0e.slice/crio-5c785fb317467101f9ff407243fb85596511c5e525db3832b38df04ed00fec73.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cebd152_5aaf_4543_be11_7ba327cfdf0e.slice/crio-conmon-883824276a85ba5f915880ccf065542294a5a3959ead2ee9e02e556f4e13a641.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cebd152_5aaf_4543_be11_7ba327cfdf0e.slice/crio-283ced1ad0dcd3181290d000eab3ac44a3033b8e32fa0ca0cd8ef03d57a7deee.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cebd152_5aaf_4543_be11_7ba327cfdf0e.slice/crio-conmon-5c785fb317467101f9ff407243fb85596511c5e525db3832b38df04ed00fec73.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84d2000a_d99f_4320_b178_874940ab7e9d.slice/crio-conmon-9799cb2e3daec429b427c098af4fc69103ad295acfa71492709948a69984306d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cebd152_5aaf_4543_be11_7ba327cfdf0e.slice/crio-883824276a85ba5f915880ccf065542294a5a3959ead2ee9e02e556f4e13a641.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cebd152_5aaf_4543_be11_7ba327cfdf0e.slice/crio-conmon-d6fc4c846f604ef5014de443285ee7dab569b41691babd5a6981585d8007df21.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cebd152_5aaf_4543_be11_7ba327cfdf0e.slice/crio-65c0b15d452f5559c24817138f1d42f00b5d08d34eff5f6a036df1ded5c70ccb.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cebd152_5aaf_4543_be11_7ba327cfdf0e.slice/crio-169e61c78b12ac121ef1306849a0fe5e05cc7125ba6f6a238a652aac3445bad8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cebd152_5aaf_4543_be11_7ba327cfdf0e.slice/crio-conmon-169e61c78b12ac121ef1306849a0fe5e05cc7125ba6f6a238a652aac3445bad8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cebd152_5aaf_4543_be11_7ba327cfdf0e.slice/crio-022023c76cc2ba5ec65353c6ec2387833901f691565a07dc55b9a6a09bf9871c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cebd152_5aaf_4543_be11_7ba327cfdf0e.slice/crio-da203f48b1e4a0c44c371939bd0a1b6ee7159680a2a3fac865d9476c3b7b9c05.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cebd152_5aaf_4543_be11_7ba327cfdf0e.slice/crio-d6fc4c846f604ef5014de443285ee7dab569b41691babd5a6981585d8007df21.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cebd152_5aaf_4543_be11_7ba327cfdf0e.slice/crio-conmon-022023c76cc2ba5ec65353c6ec2387833901f691565a07dc55b9a6a09bf9871c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cebd152_5aaf_4543_be11_7ba327cfdf0e.slice/crio-2d1cebd6bce0741d006ac48e38db4c872771c02036472a9fa71980ddfb72cef5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cebd152_5aaf_4543_be11_7ba327cfdf0e.slice/crio-conmon-2d1cebd6bce0741d006ac48e38db4c872771c02036472a9fa71980ddfb72cef5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94a95a16_6bdd_4c08_8340_13fd85455b7a.slice/crio-conmon-7f805c812aa2a019f6996c4840760eec127d8f206fe82dbfef04db5d8dadd02c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6061002c_7242_4d02_b868_6838b39f30fa.slice/crio-conmon-fd9f74a1d204e8b395d9002c1eb07158525244b64cff40fb8c86051204dcedaa.scope\": RecentStats: unable to find data in memory cache]" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.148696 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gwpkm" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.164331 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-kh9px"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.204047 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-kh9px"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.204888 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-l476l_e14738be-f83f-47ce-b0d9-ceb23d504b46/openstack-network-exporter/0.log" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.204961 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-l476l" Oct 02 10:11:12 crc kubenswrapper[4934]: E1002 10:11:12.212893 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6ab70322bf5ef2e1d2a4fd23bb06ac7d14bc671fa582c8397ceb83e14554def5" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.214282 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7214291c-3c8e-4173-a6fd-ef16170258e2-var-run\") pod \"7214291c-3c8e-4173-a6fd-ef16170258e2\" (UID: \"7214291c-3c8e-4173-a6fd-ef16170258e2\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.214418 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qj9h7\" (UniqueName: \"kubernetes.io/projected/7214291c-3c8e-4173-a6fd-ef16170258e2-kube-api-access-qj9h7\") pod \"7214291c-3c8e-4173-a6fd-ef16170258e2\" (UID: \"7214291c-3c8e-4173-a6fd-ef16170258e2\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.214446 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7214291c-3c8e-4173-a6fd-ef16170258e2-scripts\") pod \"7214291c-3c8e-4173-a6fd-ef16170258e2\" (UID: \"7214291c-3c8e-4173-a6fd-ef16170258e2\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.214481 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7214291c-3c8e-4173-a6fd-ef16170258e2-var-log-ovn\") pod \"7214291c-3c8e-4173-a6fd-ef16170258e2\" (UID: \"7214291c-3c8e-4173-a6fd-ef16170258e2\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.214512 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7214291c-3c8e-4173-a6fd-ef16170258e2-var-run-ovn\") pod \"7214291c-3c8e-4173-a6fd-ef16170258e2\" (UID: \"7214291c-3c8e-4173-a6fd-ef16170258e2\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.214626 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/7214291c-3c8e-4173-a6fd-ef16170258e2-ovn-controller-tls-certs\") pod \"7214291c-3c8e-4173-a6fd-ef16170258e2\" (UID: \"7214291c-3c8e-4173-a6fd-ef16170258e2\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.214675 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7214291c-3c8e-4173-a6fd-ef16170258e2-combined-ca-bundle\") pod \"7214291c-3c8e-4173-a6fd-ef16170258e2\" (UID: \"7214291c-3c8e-4173-a6fd-ef16170258e2\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.216054 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7214291c-3c8e-4173-a6fd-ef16170258e2-scripts" (OuterVolumeSpecName: "scripts") pod "7214291c-3c8e-4173-a6fd-ef16170258e2" (UID: "7214291c-3c8e-4173-a6fd-ef16170258e2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.216112 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7214291c-3c8e-4173-a6fd-ef16170258e2-var-run" (OuterVolumeSpecName: "var-run") pod "7214291c-3c8e-4173-a6fd-ef16170258e2" (UID: "7214291c-3c8e-4173-a6fd-ef16170258e2"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.217997 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7214291c-3c8e-4173-a6fd-ef16170258e2-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "7214291c-3c8e-4173-a6fd-ef16170258e2" (UID: "7214291c-3c8e-4173-a6fd-ef16170258e2"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.218067 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7214291c-3c8e-4173-a6fd-ef16170258e2-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "7214291c-3c8e-4173-a6fd-ef16170258e2" (UID: "7214291c-3c8e-4173-a6fd-ef16170258e2"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.222438 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.222794 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="04aff378-509c-49be-bf4c-2d59111fc910" containerName="nova-metadata-log" containerID="cri-o://0c87adf381aab6824e14a94acbf03d404203f8ec3721590d4efd316ce48e8dd8" gracePeriod=30 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.222944 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="04aff378-509c-49be-bf4c-2d59111fc910" containerName="nova-metadata-metadata" containerID="cri-o://f28c0b65a711c22b95f0ba26fc6761c9a173fc0e9b32cd123f69c153b87e25bf" gracePeriod=30 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.223252 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7214291c-3c8e-4173-a6fd-ef16170258e2-kube-api-access-qj9h7" (OuterVolumeSpecName: "kube-api-access-qj9h7") pod "7214291c-3c8e-4173-a6fd-ef16170258e2" (UID: "7214291c-3c8e-4173-a6fd-ef16170258e2"). InnerVolumeSpecName "kube-api-access-qj9h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.230248 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-ba62-account-create-76fqw"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.235889 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_9d816754-2d4c-40e6-acee-218947ab9e19/ovsdbserver-nb/0.log" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.235971 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.269710 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-ba62-account-create-76fqw"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.285019 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-65wtc" podUID="1146cda2-78fc-40c5-b0d2-f8c19368fa2c" containerName="ovs-vswitchd" containerID="cri-o://c63d866adb61766a772229b32b348b743c5300884ef0ae4924c680c6cb85c6e0" gracePeriod=28 Oct 02 10:11:12 crc kubenswrapper[4934]: E1002 10:11:12.285249 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6ab70322bf5ef2e1d2a4fd23bb06ac7d14bc671fa582c8397ceb83e14554def5" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.330732 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d816754-2d4c-40e6-acee-218947ab9e19-scripts\") pod \"9d816754-2d4c-40e6-acee-218947ab9e19\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.330795 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7tqx\" (UniqueName: \"kubernetes.io/projected/9d816754-2d4c-40e6-acee-218947ab9e19-kube-api-access-s7tqx\") pod \"9d816754-2d4c-40e6-acee-218947ab9e19\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.330824 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"9d816754-2d4c-40e6-acee-218947ab9e19\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.330947 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d816754-2d4c-40e6-acee-218947ab9e19-combined-ca-bundle\") pod \"9d816754-2d4c-40e6-acee-218947ab9e19\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.330980 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d816754-2d4c-40e6-acee-218947ab9e19-config\") pod \"9d816754-2d4c-40e6-acee-218947ab9e19\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.331022 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvqfn\" (UniqueName: \"kubernetes.io/projected/e14738be-f83f-47ce-b0d9-ceb23d504b46-kube-api-access-mvqfn\") pod \"e14738be-f83f-47ce-b0d9-ceb23d504b46\" (UID: \"e14738be-f83f-47ce-b0d9-ceb23d504b46\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.331052 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e14738be-f83f-47ce-b0d9-ceb23d504b46-ovs-rundir\") pod \"e14738be-f83f-47ce-b0d9-ceb23d504b46\" (UID: \"e14738be-f83f-47ce-b0d9-ceb23d504b46\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.331121 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e14738be-f83f-47ce-b0d9-ceb23d504b46-metrics-certs-tls-certs\") pod \"e14738be-f83f-47ce-b0d9-ceb23d504b46\" (UID: \"e14738be-f83f-47ce-b0d9-ceb23d504b46\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.331165 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e14738be-f83f-47ce-b0d9-ceb23d504b46-combined-ca-bundle\") pod \"e14738be-f83f-47ce-b0d9-ceb23d504b46\" (UID: \"e14738be-f83f-47ce-b0d9-ceb23d504b46\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.331197 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d816754-2d4c-40e6-acee-218947ab9e19-ovsdbserver-nb-tls-certs\") pod \"9d816754-2d4c-40e6-acee-218947ab9e19\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.331274 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e14738be-f83f-47ce-b0d9-ceb23d504b46-ovn-rundir\") pod \"e14738be-f83f-47ce-b0d9-ceb23d504b46\" (UID: \"e14738be-f83f-47ce-b0d9-ceb23d504b46\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.331332 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d816754-2d4c-40e6-acee-218947ab9e19-metrics-certs-tls-certs\") pod \"9d816754-2d4c-40e6-acee-218947ab9e19\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.331358 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e14738be-f83f-47ce-b0d9-ceb23d504b46-config\") pod \"e14738be-f83f-47ce-b0d9-ceb23d504b46\" (UID: \"e14738be-f83f-47ce-b0d9-ceb23d504b46\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.331375 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9d816754-2d4c-40e6-acee-218947ab9e19-ovsdb-rundir\") pod \"9d816754-2d4c-40e6-acee-218947ab9e19\" (UID: \"9d816754-2d4c-40e6-acee-218947ab9e19\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.332153 4934 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7214291c-3c8e-4173-a6fd-ef16170258e2-var-run\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.332164 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qj9h7\" (UniqueName: \"kubernetes.io/projected/7214291c-3c8e-4173-a6fd-ef16170258e2-kube-api-access-qj9h7\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.332177 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7214291c-3c8e-4173-a6fd-ef16170258e2-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.332186 4934 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7214291c-3c8e-4173-a6fd-ef16170258e2-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.332210 4934 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7214291c-3c8e-4173-a6fd-ef16170258e2-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.332938 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d816754-2d4c-40e6-acee-218947ab9e19-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "9d816754-2d4c-40e6-acee-218947ab9e19" (UID: "9d816754-2d4c-40e6-acee-218947ab9e19"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.337331 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e14738be-f83f-47ce-b0d9-ceb23d504b46-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "e14738be-f83f-47ce-b0d9-ceb23d504b46" (UID: "e14738be-f83f-47ce-b0d9-ceb23d504b46"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.341690 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 10:11:12 crc kubenswrapper[4934]: E1002 10:11:12.341917 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6ab70322bf5ef2e1d2a4fd23bb06ac7d14bc671fa582c8397ceb83e14554def5" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 02 10:11:12 crc kubenswrapper[4934]: E1002 10:11:12.341948 4934 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="05c8faa0-6986-415d-abc3-4d35291375f2" containerName="ovn-northd" Oct 02 10:11:12 crc kubenswrapper[4934]: E1002 10:11:12.342021 4934 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 02 10:11:12 crc kubenswrapper[4934]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 02 10:11:12 crc kubenswrapper[4934]: + source /usr/local/bin/container-scripts/functions Oct 02 10:11:12 crc kubenswrapper[4934]: ++ OVNBridge=br-int Oct 02 10:11:12 crc kubenswrapper[4934]: ++ OVNRemote=tcp:localhost:6642 Oct 02 10:11:12 crc kubenswrapper[4934]: ++ OVNEncapType=geneve Oct 02 10:11:12 crc kubenswrapper[4934]: ++ OVNAvailabilityZones= Oct 02 10:11:12 crc kubenswrapper[4934]: ++ EnableChassisAsGateway=true Oct 02 10:11:12 crc kubenswrapper[4934]: ++ PhysicalNetworks= Oct 02 10:11:12 crc kubenswrapper[4934]: ++ OVNHostName= Oct 02 10:11:12 crc kubenswrapper[4934]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 02 10:11:12 crc kubenswrapper[4934]: ++ ovs_dir=/var/lib/openvswitch Oct 02 10:11:12 crc kubenswrapper[4934]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 02 10:11:12 crc kubenswrapper[4934]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 02 10:11:12 crc kubenswrapper[4934]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 02 10:11:12 crc kubenswrapper[4934]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 02 10:11:12 crc kubenswrapper[4934]: + sleep 0.5 Oct 02 10:11:12 crc kubenswrapper[4934]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 02 10:11:12 crc kubenswrapper[4934]: + sleep 0.5 Oct 02 10:11:12 crc kubenswrapper[4934]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 02 10:11:12 crc kubenswrapper[4934]: + sleep 0.5 Oct 02 10:11:12 crc kubenswrapper[4934]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 02 10:11:12 crc kubenswrapper[4934]: + cleanup_ovsdb_server_semaphore Oct 02 10:11:12 crc kubenswrapper[4934]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 02 10:11:12 crc kubenswrapper[4934]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 02 10:11:12 crc kubenswrapper[4934]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-65wtc" message=< Oct 02 10:11:12 crc kubenswrapper[4934]: Exiting ovsdb-server (5) [ OK ] Oct 02 10:11:12 crc kubenswrapper[4934]: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 02 10:11:12 crc kubenswrapper[4934]: + source /usr/local/bin/container-scripts/functions Oct 02 10:11:12 crc kubenswrapper[4934]: ++ OVNBridge=br-int Oct 02 10:11:12 crc kubenswrapper[4934]: ++ OVNRemote=tcp:localhost:6642 Oct 02 10:11:12 crc kubenswrapper[4934]: ++ OVNEncapType=geneve Oct 02 10:11:12 crc kubenswrapper[4934]: ++ OVNAvailabilityZones= Oct 02 10:11:12 crc kubenswrapper[4934]: ++ EnableChassisAsGateway=true Oct 02 10:11:12 crc kubenswrapper[4934]: ++ PhysicalNetworks= Oct 02 10:11:12 crc kubenswrapper[4934]: ++ OVNHostName= Oct 02 10:11:12 crc kubenswrapper[4934]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 02 10:11:12 crc kubenswrapper[4934]: ++ ovs_dir=/var/lib/openvswitch Oct 02 10:11:12 crc kubenswrapper[4934]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 02 10:11:12 crc kubenswrapper[4934]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 02 10:11:12 crc kubenswrapper[4934]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 02 10:11:12 crc kubenswrapper[4934]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 02 10:11:12 crc kubenswrapper[4934]: + sleep 0.5 Oct 02 10:11:12 crc kubenswrapper[4934]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 02 10:11:12 crc kubenswrapper[4934]: + sleep 0.5 Oct 02 10:11:12 crc kubenswrapper[4934]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 02 10:11:12 crc kubenswrapper[4934]: + sleep 0.5 Oct 02 10:11:12 crc kubenswrapper[4934]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 02 10:11:12 crc kubenswrapper[4934]: + cleanup_ovsdb_server_semaphore Oct 02 10:11:12 crc kubenswrapper[4934]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 02 10:11:12 crc kubenswrapper[4934]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 02 10:11:12 crc kubenswrapper[4934]: > Oct 02 10:11:12 crc kubenswrapper[4934]: E1002 10:11:12.342043 4934 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 02 10:11:12 crc kubenswrapper[4934]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 02 10:11:12 crc kubenswrapper[4934]: + source /usr/local/bin/container-scripts/functions Oct 02 10:11:12 crc kubenswrapper[4934]: ++ OVNBridge=br-int Oct 02 10:11:12 crc kubenswrapper[4934]: ++ OVNRemote=tcp:localhost:6642 Oct 02 10:11:12 crc kubenswrapper[4934]: ++ OVNEncapType=geneve Oct 02 10:11:12 crc kubenswrapper[4934]: ++ OVNAvailabilityZones= Oct 02 10:11:12 crc kubenswrapper[4934]: ++ EnableChassisAsGateway=true Oct 02 10:11:12 crc kubenswrapper[4934]: ++ PhysicalNetworks= Oct 02 10:11:12 crc kubenswrapper[4934]: ++ OVNHostName= Oct 02 10:11:12 crc kubenswrapper[4934]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 02 10:11:12 crc kubenswrapper[4934]: ++ ovs_dir=/var/lib/openvswitch Oct 02 10:11:12 crc kubenswrapper[4934]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 02 10:11:12 crc kubenswrapper[4934]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 02 10:11:12 crc kubenswrapper[4934]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 02 10:11:12 crc kubenswrapper[4934]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 02 10:11:12 crc kubenswrapper[4934]: + sleep 0.5 Oct 02 10:11:12 crc kubenswrapper[4934]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 02 10:11:12 crc kubenswrapper[4934]: + sleep 0.5 Oct 02 10:11:12 crc kubenswrapper[4934]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 02 10:11:12 crc kubenswrapper[4934]: + sleep 0.5 Oct 02 10:11:12 crc kubenswrapper[4934]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 02 10:11:12 crc kubenswrapper[4934]: + cleanup_ovsdb_server_semaphore Oct 02 10:11:12 crc kubenswrapper[4934]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 02 10:11:12 crc kubenswrapper[4934]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 02 10:11:12 crc kubenswrapper[4934]: > pod="openstack/ovn-controller-ovs-65wtc" podUID="1146cda2-78fc-40c5-b0d2-f8c19368fa2c" containerName="ovsdb-server" containerID="cri-o://9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.342062 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-65wtc" podUID="1146cda2-78fc-40c5-b0d2-f8c19368fa2c" containerName="ovsdb-server" containerID="cri-o://9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892" gracePeriod=28 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.343357 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e14738be-f83f-47ce-b0d9-ceb23d504b46-config" (OuterVolumeSpecName: "config") pod "e14738be-f83f-47ce-b0d9-ceb23d504b46" (UID: "e14738be-f83f-47ce-b0d9-ceb23d504b46"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.344491 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e14738be-f83f-47ce-b0d9-ceb23d504b46-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "e14738be-f83f-47ce-b0d9-ceb23d504b46" (UID: "e14738be-f83f-47ce-b0d9-ceb23d504b46"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.346771 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "9d816754-2d4c-40e6-acee-218947ab9e19" (UID: "9d816754-2d4c-40e6-acee-218947ab9e19"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.347169 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.347357 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="06adab43-0177-440c-aeef-9386345160c4" containerName="nova-scheduler-scheduler" containerID="cri-o://b7f8b726c37d4fd8d67700b1306d05906985514bcb56092fd5f3e2779cf2dc8a" gracePeriod=30 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.348332 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d816754-2d4c-40e6-acee-218947ab9e19-config" (OuterVolumeSpecName: "config") pod "9d816754-2d4c-40e6-acee-218947ab9e19" (UID: "9d816754-2d4c-40e6-acee-218947ab9e19"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.348606 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-bfc5bc4b9-crmsr"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.348639 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d816754-2d4c-40e6-acee-218947ab9e19-scripts" (OuterVolumeSpecName: "scripts") pod "9d816754-2d4c-40e6-acee-218947ab9e19" (UID: "9d816754-2d4c-40e6-acee-218947ab9e19"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.348732 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-bfc5bc4b9-crmsr" podUID="a30daf4c-06f2-48f5-b57f-f6704fae13a9" containerName="proxy-httpd" containerID="cri-o://ba2dbd7671783785d215106bb6b057dd861995948702cd2be58811e1a9b4c38e" gracePeriod=30 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.348799 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-bfc5bc4b9-crmsr" podUID="a30daf4c-06f2-48f5-b57f-f6704fae13a9" containerName="proxy-server" containerID="cri-o://65dbab7a03568a76645932c0b29b78b90ef3f3460bb8573a3ba9cd75761a9154" gracePeriod=30 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.351621 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e14738be-f83f-47ce-b0d9-ceb23d504b46-kube-api-access-mvqfn" (OuterVolumeSpecName: "kube-api-access-mvqfn") pod "e14738be-f83f-47ce-b0d9-ceb23d504b46" (UID: "e14738be-f83f-47ce-b0d9-ceb23d504b46"). InnerVolumeSpecName "kube-api-access-mvqfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.353315 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d816754-2d4c-40e6-acee-218947ab9e19-kube-api-access-s7tqx" (OuterVolumeSpecName: "kube-api-access-s7tqx") pod "9d816754-2d4c-40e6-acee-218947ab9e19" (UID: "9d816754-2d4c-40e6-acee-218947ab9e19"). InnerVolumeSpecName "kube-api-access-s7tqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.368681 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-vzrk9"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.375882 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-vzrk9"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.389416 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.389629 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="3dda2730-9bb9-4113-80d3-168eb5905b2f" containerName="nova-cell0-conductor-conductor" containerID="cri-o://75db33a0a204cb84036a785aefe5e3a861f5bb1c02c563e44111dc7a54a02cb7" gracePeriod=30 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.400898 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.403213 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="125b40d7-516c-432f-ae51-703f2b18068d" containerName="galera" containerID="cri-o://295f3add8c46cf850cd886f2d08991a6be7f2acfd564e68d63ce13cc92ba58df" gracePeriod=30 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.435969 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.436457 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="49a440a9-1a3f-4ee9-bf28-7996fe6f222d" containerName="nova-cell1-conductor-conductor" containerID="cri-o://38289b907213ccb98a8b74c9827dd591276a04d2139306b7c6ab58968debd430" gracePeriod=30 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.437044 4934 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e14738be-f83f-47ce-b0d9-ceb23d504b46-ovs-rundir\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.437067 4934 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e14738be-f83f-47ce-b0d9-ceb23d504b46-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.437082 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e14738be-f83f-47ce-b0d9-ceb23d504b46-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.437095 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9d816754-2d4c-40e6-acee-218947ab9e19-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.437105 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d816754-2d4c-40e6-acee-218947ab9e19-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.437114 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7tqx\" (UniqueName: \"kubernetes.io/projected/9d816754-2d4c-40e6-acee-218947ab9e19-kube-api-access-s7tqx\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.437145 4934 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.437161 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d816754-2d4c-40e6-acee-218947ab9e19-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.437170 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvqfn\" (UniqueName: \"kubernetes.io/projected/e14738be-f83f-47ce-b0d9-ceb23d504b46-kube-api-access-mvqfn\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.441923 4934 generic.go:334] "Generic (PLEG): container finished" podID="ecc76715-8b05-4529-bd7b-289d7f32eff5" containerID="4d19c2050c253cd17d4e337c977a53e89370af2b71d1d6898f8b9f9d88a2f2b2" exitCode=143 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.442049 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-649ccd9666-dh5rs" event={"ID":"ecc76715-8b05-4529-bd7b-289d7f32eff5","Type":"ContainerDied","Data":"4d19c2050c253cd17d4e337c977a53e89370af2b71d1d6898f8b9f9d88a2f2b2"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.448478 4934 generic.go:334] "Generic (PLEG): container finished" podID="547c6c6d-0f01-4feb-acc6-2e3045407b64" containerID="ca5c961136b99e80a8ae0bb9c203d870c7355071cc4a476bdb7c41606bdef8b7" exitCode=143 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.448600 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6bf48f9d67-g95nd" event={"ID":"547c6c6d-0f01-4feb-acc6-2e3045407b64","Type":"ContainerDied","Data":"ca5c961136b99e80a8ae0bb9c203d870c7355071cc4a476bdb7c41606bdef8b7"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.454080 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6c8h9"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.469478 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6c8h9"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.469862 4934 generic.go:334] "Generic (PLEG): container finished" podID="6d79c694-2253-419e-b63b-4a38884ac57e" containerID="959724f7a775ac28fa7ae269f3be79b301f1332c985d6f8a80a47def488f67f5" exitCode=0 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.469974 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-57456ffd97-rqc8b" event={"ID":"6d79c694-2253-419e-b63b-4a38884ac57e","Type":"ContainerDied","Data":"959724f7a775ac28fa7ae269f3be79b301f1332c985d6f8a80a47def488f67f5"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.497107 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.497402 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://b9baba9779ef141c5b9d9cb5a3a64ef9893f0e264bbcdb5320d4acc16116a004" gracePeriod=30 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.506466 4934 generic.go:334] "Generic (PLEG): container finished" podID="d15b0450-e100-4ee7-ad62-88601eedc4f2" containerID="c83811ac76f059e3b264882bb3a6859bb99aa9ea06694e42d9967733e316ac15" exitCode=143 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.506706 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-b97fc8746-vlx9z" event={"ID":"d15b0450-e100-4ee7-ad62-88601eedc4f2","Type":"ContainerDied","Data":"c83811ac76f059e3b264882bb3a6859bb99aa9ea06694e42d9967733e316ac15"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.521880 4934 generic.go:334] "Generic (PLEG): container finished" podID="9165cca6-da95-45ae-87c3-b2829756db3b" containerID="9dde5aaa3e65635d53137792cb4e654aa61c584c4bc197377b96acd4ff67091e" exitCode=143 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.522119 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9165cca6-da95-45ae-87c3-b2829756db3b","Type":"ContainerDied","Data":"9dde5aaa3e65635d53137792cb4e654aa61c584c4bc197377b96acd4ff67091e"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.538286 4934 generic.go:334] "Generic (PLEG): container finished" podID="9c99e264-05b3-4cd7-91ae-cabb6d10466a" containerID="7e1c034adb12673de1bc4203a0717c7690b9b6165dd60bea908b057b397ba31e" exitCode=137 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.538564 4934 scope.go:117] "RemoveContainer" containerID="7e1c034adb12673de1bc4203a0717c7690b9b6165dd60bea908b057b397ba31e" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.538809 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.545189 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96nfv\" (UniqueName: \"kubernetes.io/projected/9c99e264-05b3-4cd7-91ae-cabb6d10466a-kube-api-access-96nfv\") pod \"9c99e264-05b3-4cd7-91ae-cabb6d10466a\" (UID: \"9c99e264-05b3-4cd7-91ae-cabb6d10466a\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.548745 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9c99e264-05b3-4cd7-91ae-cabb6d10466a-openstack-config\") pod \"9c99e264-05b3-4cd7-91ae-cabb6d10466a\" (UID: \"9c99e264-05b3-4cd7-91ae-cabb6d10466a\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.548823 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c99e264-05b3-4cd7-91ae-cabb6d10466a-combined-ca-bundle\") pod \"9c99e264-05b3-4cd7-91ae-cabb6d10466a\" (UID: \"9c99e264-05b3-4cd7-91ae-cabb6d10466a\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.548895 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9c99e264-05b3-4cd7-91ae-cabb6d10466a-openstack-config-secret\") pod \"9c99e264-05b3-4cd7-91ae-cabb6d10466a\" (UID: \"9c99e264-05b3-4cd7-91ae-cabb6d10466a\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.571312 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_6061002c-7242-4d02-b868-6838b39f30fa/ovsdbserver-sb/0.log" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.571602 4934 generic.go:334] "Generic (PLEG): container finished" podID="6061002c-7242-4d02-b868-6838b39f30fa" containerID="abc6a88d34ca1719e3b11b44524da713532a93344a46bb1626798816c0caa7a2" exitCode=143 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.571696 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6061002c-7242-4d02-b868-6838b39f30fa","Type":"ContainerDied","Data":"abc6a88d34ca1719e3b11b44524da713532a93344a46bb1626798816c0caa7a2"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.576055 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c99e264-05b3-4cd7-91ae-cabb6d10466a-kube-api-access-96nfv" (OuterVolumeSpecName: "kube-api-access-96nfv") pod "9c99e264-05b3-4cd7-91ae-cabb6d10466a" (UID: "9c99e264-05b3-4cd7-91ae-cabb6d10466a"). InnerVolumeSpecName "kube-api-access-96nfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.587121 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.599777 4934 generic.go:334] "Generic (PLEG): container finished" podID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerID="9373e5fbc3b166030467efc3d48ed6860fef2654fa4d21f813440b32765b3063" exitCode=0 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.599826 4934 generic.go:334] "Generic (PLEG): container finished" podID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerID="2d1cebd6bce0741d006ac48e38db4c872771c02036472a9fa71980ddfb72cef5" exitCode=0 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.599835 4934 generic.go:334] "Generic (PLEG): container finished" podID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerID="6b7e46bf6361363ea3ce0effc8f79e7c186a4d64e12c8053c444d0f89313722e" exitCode=0 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.599842 4934 generic.go:334] "Generic (PLEG): container finished" podID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerID="169e61c78b12ac121ef1306849a0fe5e05cc7125ba6f6a238a652aac3445bad8" exitCode=0 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.599849 4934 generic.go:334] "Generic (PLEG): container finished" podID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerID="d6fc4c846f604ef5014de443285ee7dab569b41691babd5a6981585d8007df21" exitCode=0 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.599857 4934 generic.go:334] "Generic (PLEG): container finished" podID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerID="1e2427a94989f9a77ca59046a739fc7849881116c29e79c766ddc52debf75ec3" exitCode=0 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.599862 4934 generic.go:334] "Generic (PLEG): container finished" podID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerID="883824276a85ba5f915880ccf065542294a5a3959ead2ee9e02e556f4e13a641" exitCode=0 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.600012 4934 generic.go:334] "Generic (PLEG): container finished" podID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerID="022023c76cc2ba5ec65353c6ec2387833901f691565a07dc55b9a6a09bf9871c" exitCode=0 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.600021 4934 generic.go:334] "Generic (PLEG): container finished" podID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerID="da203f48b1e4a0c44c371939bd0a1b6ee7159680a2a3fac865d9476c3b7b9c05" exitCode=0 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.600172 4934 generic.go:334] "Generic (PLEG): container finished" podID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerID="c0f5815d02eff4beefc969cb8468e9b85a60d871f99cc29f33aa27f2cd50039f" exitCode=0 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.600184 4934 generic.go:334] "Generic (PLEG): container finished" podID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerID="65c0b15d452f5559c24817138f1d42f00b5d08d34eff5f6a036df1ded5c70ccb" exitCode=0 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.600190 4934 generic.go:334] "Generic (PLEG): container finished" podID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerID="5c785fb317467101f9ff407243fb85596511c5e525db3832b38df04ed00fec73" exitCode=0 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.600196 4934 generic.go:334] "Generic (PLEG): container finished" podID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerID="283ced1ad0dcd3181290d000eab3ac44a3033b8e32fa0ca0cd8ef03d57a7deee" exitCode=0 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.600202 4934 generic.go:334] "Generic (PLEG): container finished" podID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerID="478b442a459773165d46dac69a02032f32c9b37d6398eff28160ced5dd0ee4b4" exitCode=0 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.600383 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerDied","Data":"9373e5fbc3b166030467efc3d48ed6860fef2654fa4d21f813440b32765b3063"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.600410 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerDied","Data":"2d1cebd6bce0741d006ac48e38db4c872771c02036472a9fa71980ddfb72cef5"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.600421 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerDied","Data":"6b7e46bf6361363ea3ce0effc8f79e7c186a4d64e12c8053c444d0f89313722e"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.600563 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerDied","Data":"169e61c78b12ac121ef1306849a0fe5e05cc7125ba6f6a238a652aac3445bad8"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.600596 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerDied","Data":"d6fc4c846f604ef5014de443285ee7dab569b41691babd5a6981585d8007df21"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.600608 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerDied","Data":"1e2427a94989f9a77ca59046a739fc7849881116c29e79c766ddc52debf75ec3"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.600616 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerDied","Data":"883824276a85ba5f915880ccf065542294a5a3959ead2ee9e02e556f4e13a641"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.600625 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerDied","Data":"022023c76cc2ba5ec65353c6ec2387833901f691565a07dc55b9a6a09bf9871c"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.600635 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerDied","Data":"da203f48b1e4a0c44c371939bd0a1b6ee7159680a2a3fac865d9476c3b7b9c05"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.600643 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerDied","Data":"c0f5815d02eff4beefc969cb8468e9b85a60d871f99cc29f33aa27f2cd50039f"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.600651 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerDied","Data":"65c0b15d452f5559c24817138f1d42f00b5d08d34eff5f6a036df1ded5c70ccb"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.600833 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerDied","Data":"5c785fb317467101f9ff407243fb85596511c5e525db3832b38df04ed00fec73"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.600847 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerDied","Data":"283ced1ad0dcd3181290d000eab3ac44a3033b8e32fa0ca0cd8ef03d57a7deee"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.600856 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerDied","Data":"478b442a459773165d46dac69a02032f32c9b37d6398eff28160ced5dd0ee4b4"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.614121 4934 generic.go:334] "Generic (PLEG): container finished" podID="cd41dc17-6659-4b43-9758-f67d831993ba" containerID="ddc003d07d24a576f93d547d9f6f180b883faaf22fef13259bf076d508867ebe" exitCode=143 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.614201 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cd41dc17-6659-4b43-9758-f67d831993ba","Type":"ContainerDied","Data":"ddc003d07d24a576f93d547d9f6f180b883faaf22fef13259bf076d508867ebe"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.622729 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c99e264-05b3-4cd7-91ae-cabb6d10466a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c99e264-05b3-4cd7-91ae-cabb6d10466a" (UID: "9c99e264-05b3-4cd7-91ae-cabb6d10466a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.623008 4934 scope.go:117] "RemoveContainer" containerID="7e1c034adb12673de1bc4203a0717c7690b9b6165dd60bea908b057b397ba31e" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.625836 4934 generic.go:334] "Generic (PLEG): container finished" podID="350928d4-5a0b-4331-b1c2-220517d313a5" containerID="9245f97a19f447f9859a60a1c0d5f3f95ccff037fb5235ac35a03f447f11628f" exitCode=143 Oct 02 10:11:12 crc kubenswrapper[4934]: E1002 10:11:12.625857 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e1c034adb12673de1bc4203a0717c7690b9b6165dd60bea908b057b397ba31e\": container with ID starting with 7e1c034adb12673de1bc4203a0717c7690b9b6165dd60bea908b057b397ba31e not found: ID does not exist" containerID="7e1c034adb12673de1bc4203a0717c7690b9b6165dd60bea908b057b397ba31e" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.625883 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e1c034adb12673de1bc4203a0717c7690b9b6165dd60bea908b057b397ba31e"} err="failed to get container status \"7e1c034adb12673de1bc4203a0717c7690b9b6165dd60bea908b057b397ba31e\": rpc error: code = NotFound desc = could not find container \"7e1c034adb12673de1bc4203a0717c7690b9b6165dd60bea908b057b397ba31e\": container with ID starting with 7e1c034adb12673de1bc4203a0717c7690b9b6165dd60bea908b057b397ba31e not found: ID does not exist" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.625912 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8648b56f4-5f8xk" event={"ID":"350928d4-5a0b-4331-b1c2-220517d313a5","Type":"ContainerDied","Data":"9245f97a19f447f9859a60a1c0d5f3f95ccff037fb5235ac35a03f447f11628f"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.638309 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_9d816754-2d4c-40e6-acee-218947ab9e19/ovsdbserver-nb/0.log" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.638379 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"9d816754-2d4c-40e6-acee-218947ab9e19","Type":"ContainerDied","Data":"3f2ad5d6f4972b1e2bcb07aa5ec6d1f71a859bef238a2c99f2a01002f7a2267b"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.638415 4934 scope.go:117] "RemoveContainer" containerID="66355c51e4027fd1c4a3713d064cb84130067b07985283b47ad3d5d705b2d508" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.638522 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.664204 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-dns-swift-storage-0\") pod \"20b39d07-1559-4d11-b9ff-6ef97142a58a\" (UID: \"20b39d07-1559-4d11-b9ff-6ef97142a58a\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.664270 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-ovsdbserver-sb\") pod \"20b39d07-1559-4d11-b9ff-6ef97142a58a\" (UID: \"20b39d07-1559-4d11-b9ff-6ef97142a58a\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.664375 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-config\") pod \"20b39d07-1559-4d11-b9ff-6ef97142a58a\" (UID: \"20b39d07-1559-4d11-b9ff-6ef97142a58a\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.664536 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65775\" (UniqueName: \"kubernetes.io/projected/20b39d07-1559-4d11-b9ff-6ef97142a58a-kube-api-access-65775\") pod \"20b39d07-1559-4d11-b9ff-6ef97142a58a\" (UID: \"20b39d07-1559-4d11-b9ff-6ef97142a58a\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.664781 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-ovsdbserver-nb\") pod \"20b39d07-1559-4d11-b9ff-6ef97142a58a\" (UID: \"20b39d07-1559-4d11-b9ff-6ef97142a58a\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.664812 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-dns-svc\") pod \"20b39d07-1559-4d11-b9ff-6ef97142a58a\" (UID: \"20b39d07-1559-4d11-b9ff-6ef97142a58a\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.665186 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96nfv\" (UniqueName: \"kubernetes.io/projected/9c99e264-05b3-4cd7-91ae-cabb6d10466a-kube-api-access-96nfv\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.665201 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c99e264-05b3-4cd7-91ae-cabb6d10466a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.671644 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="97560469-c459-48f4-901e-54eaded32537" containerName="rabbitmq" containerID="cri-o://2432f321bd3453c2b78b0dac285ee73ee7b8165fefef31515a2963451230f5a8" gracePeriod=604800 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.678075 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance82a2-account-delete-sv9vf"] Oct 02 10:11:12 crc kubenswrapper[4934]: E1002 10:11:12.703167 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="75db33a0a204cb84036a785aefe5e3a861f5bb1c02c563e44111dc7a54a02cb7" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.712127 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b39d07-1559-4d11-b9ff-6ef97142a58a-kube-api-access-65775" (OuterVolumeSpecName: "kube-api-access-65775") pod "20b39d07-1559-4d11-b9ff-6ef97142a58a" (UID: "20b39d07-1559-4d11-b9ff-6ef97142a58a"). InnerVolumeSpecName "kube-api-access-65775". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.712824 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c99e264-05b3-4cd7-91ae-cabb6d10466a-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "9c99e264-05b3-4cd7-91ae-cabb6d10466a" (UID: "9c99e264-05b3-4cd7-91ae-cabb6d10466a"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:12 crc kubenswrapper[4934]: E1002 10:11:12.714828 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="75db33a0a204cb84036a785aefe5e3a861f5bb1c02c563e44111dc7a54a02cb7" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.716737 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gwpkm" event={"ID":"7214291c-3c8e-4173-a6fd-ef16170258e2","Type":"ContainerDied","Data":"5927bcbc36ec9b1bb775b2c5dde737fd22af01358495cd76bafa0b0d039a25cc"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.716880 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gwpkm" Oct 02 10:11:12 crc kubenswrapper[4934]: E1002 10:11:12.717360 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="75db33a0a204cb84036a785aefe5e3a861f5bb1c02c563e44111dc7a54a02cb7" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 02 10:11:12 crc kubenswrapper[4934]: E1002 10:11:12.717421 4934 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="3dda2730-9bb9-4113-80d3-168eb5905b2f" containerName="nova-cell0-conductor-conductor" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.726077 4934 scope.go:117] "RemoveContainer" containerID="be30a7274c35eb6dd2db94c2805716e312b3406c00793273d691971991038504" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.726546 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_6061002c-7242-4d02-b868-6838b39f30fa/ovsdbserver-sb/0.log" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.726634 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.734697 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-l476l_e14738be-f83f-47ce-b0d9-ceb23d504b46/openstack-network-exporter/0.log" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.734772 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-l476l" event={"ID":"e14738be-f83f-47ce-b0d9-ceb23d504b46","Type":"ContainerDied","Data":"02a4994e4aa0fe60ba80ae18379efa8ae4e5356c97b4752a007e79f341858e26"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.735004 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-l476l" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.736008 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d816754-2d4c-40e6-acee-218947ab9e19-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "9d816754-2d4c-40e6-acee-218947ab9e19" (UID: "9d816754-2d4c-40e6-acee-218947ab9e19"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.767020 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"6061002c-7242-4d02-b868-6838b39f30fa\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.767054 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6061002c-7242-4d02-b868-6838b39f30fa-metrics-certs-tls-certs\") pod \"6061002c-7242-4d02-b868-6838b39f30fa\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.767094 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6061002c-7242-4d02-b868-6838b39f30fa-config\") pod \"6061002c-7242-4d02-b868-6838b39f30fa\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.767134 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6061002c-7242-4d02-b868-6838b39f30fa-scripts\") pod \"6061002c-7242-4d02-b868-6838b39f30fa\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.767174 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6061002c-7242-4d02-b868-6838b39f30fa-combined-ca-bundle\") pod \"6061002c-7242-4d02-b868-6838b39f30fa\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.767247 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6061002c-7242-4d02-b868-6838b39f30fa-ovsdb-rundir\") pod \"6061002c-7242-4d02-b868-6838b39f30fa\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.767320 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6061002c-7242-4d02-b868-6838b39f30fa-ovsdbserver-sb-tls-certs\") pod \"6061002c-7242-4d02-b868-6838b39f30fa\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.767479 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfkgn\" (UniqueName: \"kubernetes.io/projected/6061002c-7242-4d02-b868-6838b39f30fa-kube-api-access-rfkgn\") pod \"6061002c-7242-4d02-b868-6838b39f30fa\" (UID: \"6061002c-7242-4d02-b868-6838b39f30fa\") " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.768844 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6061002c-7242-4d02-b868-6838b39f30fa-config" (OuterVolumeSpecName: "config") pod "6061002c-7242-4d02-b868-6838b39f30fa" (UID: "6061002c-7242-4d02-b868-6838b39f30fa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.769131 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6061002c-7242-4d02-b868-6838b39f30fa-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "6061002c-7242-4d02-b868-6838b39f30fa" (UID: "6061002c-7242-4d02-b868-6838b39f30fa"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.769475 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7214291c-3c8e-4173-a6fd-ef16170258e2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7214291c-3c8e-4173-a6fd-ef16170258e2" (UID: "7214291c-3c8e-4173-a6fd-ef16170258e2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.770920 4934 generic.go:334] "Generic (PLEG): container finished" podID="20b39d07-1559-4d11-b9ff-6ef97142a58a" containerID="aac633fecdf45004f95036aeac08948cfcfe75a04ef1942aef60308cc2356517" exitCode=0 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.770991 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" event={"ID":"20b39d07-1559-4d11-b9ff-6ef97142a58a","Type":"ContainerDied","Data":"aac633fecdf45004f95036aeac08948cfcfe75a04ef1942aef60308cc2356517"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.771076 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.772044 4934 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d816754-2d4c-40e6-acee-218947ab9e19-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.772066 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7214291c-3c8e-4173-a6fd-ef16170258e2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.772075 4934 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9c99e264-05b3-4cd7-91ae-cabb6d10466a-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.772088 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6061002c-7242-4d02-b868-6838b39f30fa-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.772102 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65775\" (UniqueName: \"kubernetes.io/projected/20b39d07-1559-4d11-b9ff-6ef97142a58a-kube-api-access-65775\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.772110 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6061002c-7242-4d02-b868-6838b39f30fa-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.776450 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6061002c-7242-4d02-b868-6838b39f30fa-scripts" (OuterVolumeSpecName: "scripts") pod "6061002c-7242-4d02-b868-6838b39f30fa" (UID: "6061002c-7242-4d02-b868-6838b39f30fa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.779481 4934 generic.go:334] "Generic (PLEG): container finished" podID="94a95a16-6bdd-4c08-8340-13fd85455b7a" containerID="7f805c812aa2a019f6996c4840760eec127d8f206fe82dbfef04db5d8dadd02c" exitCode=0 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.779685 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"94a95a16-6bdd-4c08-8340-13fd85455b7a","Type":"ContainerDied","Data":"7f805c812aa2a019f6996c4840760eec127d8f206fe82dbfef04db5d8dadd02c"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.782640 4934 generic.go:334] "Generic (PLEG): container finished" podID="d4709e8f-24b0-48e6-ab94-39ec23f9ff64" containerID="b19b0bf49673fda59a55fb96cbe24384993711b894ffa56c5b3e5b90372c5231" exitCode=143 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.782699 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d4709e8f-24b0-48e6-ab94-39ec23f9ff64","Type":"ContainerDied","Data":"b19b0bf49673fda59a55fb96cbe24384993711b894ffa56c5b3e5b90372c5231"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.785115 4934 generic.go:334] "Generic (PLEG): container finished" podID="84d2000a-d99f-4320-b178-874940ab7e9d" containerID="9799cb2e3daec429b427c098af4fc69103ad295acfa71492709948a69984306d" exitCode=143 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.785196 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"84d2000a-d99f-4320-b178-874940ab7e9d","Type":"ContainerDied","Data":"9799cb2e3daec429b427c098af4fc69103ad295acfa71492709948a69984306d"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.794700 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"https://10.217.0.197:6080/vnc_lite.html\": dial tcp 10.217.0.197:6080: connect: connection refused" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.802909 4934 generic.go:334] "Generic (PLEG): container finished" podID="1146cda2-78fc-40c5-b0d2-f8c19368fa2c" containerID="9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892" exitCode=0 Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.802954 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-65wtc" event={"ID":"1146cda2-78fc-40c5-b0d2-f8c19368fa2c","Type":"ContainerDied","Data":"9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892"} Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.815088 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "6061002c-7242-4d02-b868-6838b39f30fa" (UID: "6061002c-7242-4d02-b868-6838b39f30fa"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.816717 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6061002c-7242-4d02-b868-6838b39f30fa-kube-api-access-rfkgn" (OuterVolumeSpecName: "kube-api-access-rfkgn") pod "6061002c-7242-4d02-b868-6838b39f30fa" (UID: "6061002c-7242-4d02-b868-6838b39f30fa"). InnerVolumeSpecName "kube-api-access-rfkgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.838775 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder36fe-account-delete-n2bk5"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.851715 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e14738be-f83f-47ce-b0d9-ceb23d504b46-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e14738be-f83f-47ce-b0d9-ceb23d504b46" (UID: "e14738be-f83f-47ce-b0d9-ceb23d504b46"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.874705 4934 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.874733 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6061002c-7242-4d02-b868-6838b39f30fa-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.874742 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e14738be-f83f-47ce-b0d9-ceb23d504b46-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.874753 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfkgn\" (UniqueName: \"kubernetes.io/projected/6061002c-7242-4d02-b868-6838b39f30fa-kube-api-access-rfkgn\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.910570 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapi26a6-account-delete-ng2mb"] Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.933740 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "20b39d07-1559-4d11-b9ff-6ef97142a58a" (UID: "20b39d07-1559-4d11-b9ff-6ef97142a58a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.956950 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09b1ea6c-4d4a-427d-a66e-e3db8b198316" path="/var/lib/kubelet/pods/09b1ea6c-4d4a-427d-a66e-e3db8b198316/volumes" Oct 02 10:11:12 crc kubenswrapper[4934]: I1002 10:11:12.980427 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6061002c-7242-4d02-b868-6838b39f30fa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6061002c-7242-4d02-b868-6838b39f30fa" (UID: "6061002c-7242-4d02-b868-6838b39f30fa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.006825 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.006849 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c63f790-0724-4a40-8b22-9895eb3b7f8b" path="/var/lib/kubelet/pods/1c63f790-0724-4a40-8b22-9895eb3b7f8b/volumes" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.012158 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26f3aea9-8e30-4ec3-b440-c00fe8f624da" path="/var/lib/kubelet/pods/26f3aea9-8e30-4ec3-b440-c00fe8f624da/volumes" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.019041 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30903f5b-46ed-4b59-ad1e-073466cac566" path="/var/lib/kubelet/pods/30903f5b-46ed-4b59-ad1e-073466cac566/volumes" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.019871 4934 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.020683 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3640bb49-1f9b-4bfd-9f11-4e6f5662bf76" path="/var/lib/kubelet/pods/3640bb49-1f9b-4bfd-9f11-4e6f5662bf76/volumes" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.021276 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d01321b-b929-4eb8-a780-7a4b4adae54f" path="/var/lib/kubelet/pods/3d01321b-b929-4eb8-a780-7a4b4adae54f/volumes" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.021970 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46721f65-4fa6-4059-aeb2-c9e62ed22cb1" path="/var/lib/kubelet/pods/46721f65-4fa6-4059-aeb2-c9e62ed22cb1/volumes" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.022347 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "20b39d07-1559-4d11-b9ff-6ef97142a58a" (UID: "20b39d07-1559-4d11-b9ff-6ef97142a58a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.025631 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="490173ef-a363-4b84-8802-6a69bb9bdb76" path="/var/lib/kubelet/pods/490173ef-a363-4b84-8802-6a69bb9bdb76/volumes" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.026596 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a357498-07a9-439b-9086-b431f165ee7e" path="/var/lib/kubelet/pods/4a357498-07a9-439b-9086-b431f165ee7e/volumes" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.027733 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52c52806-3d2a-4fac-938d-ba4a58ba77d5" path="/var/lib/kubelet/pods/52c52806-3d2a-4fac-938d-ba4a58ba77d5/volumes" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.027813 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7214291c-3c8e-4173-a6fd-ef16170258e2-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "7214291c-3c8e-4173-a6fd-ef16170258e2" (UID: "7214291c-3c8e-4173-a6fd-ef16170258e2"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.030532 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8841e9fd-caad-4ebd-8a89-659e10a0505b" path="/var/lib/kubelet/pods/8841e9fd-caad-4ebd-8a89-659e10a0505b/volumes" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.033358 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99a3a542-8f10-4571-9424-dee0602b290e" path="/var/lib/kubelet/pods/99a3a542-8f10-4571-9424-dee0602b290e/volumes" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.034696 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1e933bf-e455-43c2-a996-b05a3bca3c96" path="/var/lib/kubelet/pods/a1e933bf-e455-43c2-a996-b05a3bca3c96/volumes" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.035452 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae0d619d-6fbb-4758-80bc-95bdfe6f2022" path="/var/lib/kubelet/pods/ae0d619d-6fbb-4758-80bc-95bdfe6f2022/volumes" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.036884 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b" path="/var/lib/kubelet/pods/bb82cdce-d0b5-48f8-bd1b-d3467fa5a84b/volumes" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.039423 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be42c16a-682c-44b2-a53a-d33085d3a681" path="/var/lib/kubelet/pods/be42c16a-682c-44b2-a53a-d33085d3a681/volumes" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.039964 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed820a81-72fc-4d84-876f-9def630d4ff7" path="/var/lib/kubelet/pods/ed820a81-72fc-4d84-876f-9def630d4ff7/volumes" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.043052 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f591dece-fc9a-430d-b0b4-5de71b477541" path="/var/lib/kubelet/pods/f591dece-fc9a-430d-b0b4-5de71b477541/volumes" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.043513 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe286dcd-593e-48ee-8140-aca585425610" path="/var/lib/kubelet/pods/fe286dcd-593e-48ee-8140-aca585425610/volumes" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.070785 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d816754-2d4c-40e6-acee-218947ab9e19-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d816754-2d4c-40e6-acee-218947ab9e19" (UID: "9d816754-2d4c-40e6-acee-218947ab9e19"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.112839 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.112868 4934 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/7214291c-3c8e-4173-a6fd-ef16170258e2-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.112878 4934 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.112888 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6061002c-7242-4d02-b868-6838b39f30fa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.112897 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d816754-2d4c-40e6-acee-218947ab9e19-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.113174 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6061002c-7242-4d02-b868-6838b39f30fa-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "6061002c-7242-4d02-b868-6838b39f30fa" (UID: "6061002c-7242-4d02-b868-6838b39f30fa"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.136396 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "20b39d07-1559-4d11-b9ff-6ef97142a58a" (UID: "20b39d07-1559-4d11-b9ff-6ef97142a58a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.146549 4934 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.183218 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "20b39d07-1559-4d11-b9ff-6ef97142a58a" (UID: "20b39d07-1559-4d11-b9ff-6ef97142a58a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.211979 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c99e264-05b3-4cd7-91ae-cabb6d10466a-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "9c99e264-05b3-4cd7-91ae-cabb6d10466a" (UID: "9c99e264-05b3-4cd7-91ae-cabb6d10466a"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.214764 4934 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6061002c-7242-4d02-b868-6838b39f30fa-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.214796 4934 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.214809 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.214821 4934 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9c99e264-05b3-4cd7-91ae-cabb6d10466a-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.214834 4934 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:13 crc kubenswrapper[4934]: E1002 10:11:13.214929 4934 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 02 10:11:13 crc kubenswrapper[4934]: E1002 10:11:13.215015 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d2b6e083-a523-48e2-8157-088036174e12-config-data podName:d2b6e083-a523-48e2-8157-088036174e12 nodeName:}" failed. No retries permitted until 2025-10-02 10:11:17.214997683 +0000 UTC m=+1348.967639205 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/d2b6e083-a523-48e2-8157-088036174e12-config-data") pod "rabbitmq-cell1-server-0" (UID: "d2b6e083-a523-48e2-8157-088036174e12") : configmap "rabbitmq-cell1-config-data" not found Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.215796 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-config" (OuterVolumeSpecName: "config") pod "20b39d07-1559-4d11-b9ff-6ef97142a58a" (UID: "20b39d07-1559-4d11-b9ff-6ef97142a58a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.247815 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6061002c-7242-4d02-b868-6838b39f30fa-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "6061002c-7242-4d02-b868-6838b39f30fa" (UID: "6061002c-7242-4d02-b868-6838b39f30fa"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.249177 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e14738be-f83f-47ce-b0d9-ceb23d504b46-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "e14738be-f83f-47ce-b0d9-ceb23d504b46" (UID: "e14738be-f83f-47ce-b0d9-ceb23d504b46"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.250356 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d816754-2d4c-40e6-acee-218947ab9e19-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "9d816754-2d4c-40e6-acee-218947ab9e19" (UID: "9d816754-2d4c-40e6-acee-218947ab9e19"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.320527 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6061002c-7242-4d02-b868-6838b39f30fa-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.320562 4934 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e14738be-f83f-47ce-b0d9-ceb23d504b46-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.320571 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20b39d07-1559-4d11-b9ff-6ef97142a58a-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.320602 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d816754-2d4c-40e6-acee-218947ab9e19-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.498268 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement5c1f-account-delete-2jjrd"] Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.560049 4934 scope.go:117] "RemoveContainer" containerID="5ab65d1f2a5fdc33711875776226a728389485f48a057c73387e2db56049cbbc" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.616374 4934 scope.go:117] "RemoveContainer" containerID="336b71c58bd0e5981fd64408bac6c360461733df6dbdbba45a423f97c6077a34" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.710072 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6bf48f9d67-g95nd" Oct 02 10:11:13 crc kubenswrapper[4934]: E1002 10:11:13.730147 4934 projected.go:263] Couldn't get secret openstack/swift-proxy-config-data: secret "swift-proxy-config-data" not found Oct 02 10:11:13 crc kubenswrapper[4934]: E1002 10:11:13.730207 4934 projected.go:263] Couldn't get secret openstack/swift-conf: secret "swift-conf" not found Oct 02 10:11:13 crc kubenswrapper[4934]: E1002 10:11:13.730222 4934 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 10:11:13 crc kubenswrapper[4934]: E1002 10:11:13.730257 4934 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-proxy-bfc5bc4b9-crmsr: [secret "swift-proxy-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Oct 02 10:11:13 crc kubenswrapper[4934]: E1002 10:11:13.730318 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a30daf4c-06f2-48f5-b57f-f6704fae13a9-etc-swift podName:a30daf4c-06f2-48f5-b57f-f6704fae13a9 nodeName:}" failed. No retries permitted until 2025-10-02 10:11:17.730299308 +0000 UTC m=+1349.482940830 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a30daf4c-06f2-48f5-b57f-f6704fae13a9-etc-swift") pod "swift-proxy-bfc5bc4b9-crmsr" (UID: "a30daf4c-06f2-48f5-b57f-f6704fae13a9") : [secret "swift-proxy-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.831022 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxzdw\" (UniqueName: \"kubernetes.io/projected/547c6c6d-0f01-4feb-acc6-2e3045407b64-kube-api-access-vxzdw\") pod \"547c6c6d-0f01-4feb-acc6-2e3045407b64\" (UID: \"547c6c6d-0f01-4feb-acc6-2e3045407b64\") " Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.831116 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/547c6c6d-0f01-4feb-acc6-2e3045407b64-config-data\") pod \"547c6c6d-0f01-4feb-acc6-2e3045407b64\" (UID: \"547c6c6d-0f01-4feb-acc6-2e3045407b64\") " Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.831207 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/547c6c6d-0f01-4feb-acc6-2e3045407b64-combined-ca-bundle\") pod \"547c6c6d-0f01-4feb-acc6-2e3045407b64\" (UID: \"547c6c6d-0f01-4feb-acc6-2e3045407b64\") " Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.831237 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/547c6c6d-0f01-4feb-acc6-2e3045407b64-config-data-custom\") pod \"547c6c6d-0f01-4feb-acc6-2e3045407b64\" (UID: \"547c6c6d-0f01-4feb-acc6-2e3045407b64\") " Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.831299 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/547c6c6d-0f01-4feb-acc6-2e3045407b64-logs\") pod \"547c6c6d-0f01-4feb-acc6-2e3045407b64\" (UID: \"547c6c6d-0f01-4feb-acc6-2e3045407b64\") " Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.832323 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/547c6c6d-0f01-4feb-acc6-2e3045407b64-logs" (OuterVolumeSpecName: "logs") pod "547c6c6d-0f01-4feb-acc6-2e3045407b64" (UID: "547c6c6d-0f01-4feb-acc6-2e3045407b64"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.837449 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/547c6c6d-0f01-4feb-acc6-2e3045407b64-kube-api-access-vxzdw" (OuterVolumeSpecName: "kube-api-access-vxzdw") pod "547c6c6d-0f01-4feb-acc6-2e3045407b64" (UID: "547c6c6d-0f01-4feb-acc6-2e3045407b64"). InnerVolumeSpecName "kube-api-access-vxzdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.839831 4934 generic.go:334] "Generic (PLEG): container finished" podID="49a440a9-1a3f-4ee9-bf28-7996fe6f222d" containerID="38289b907213ccb98a8b74c9827dd591276a04d2139306b7c6ab58968debd430" exitCode=0 Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.839948 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"49a440a9-1a3f-4ee9-bf28-7996fe6f222d","Type":"ContainerDied","Data":"38289b907213ccb98a8b74c9827dd591276a04d2139306b7c6ab58968debd430"} Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.841894 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/547c6c6d-0f01-4feb-acc6-2e3045407b64-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "547c6c6d-0f01-4feb-acc6-2e3045407b64" (UID: "547c6c6d-0f01-4feb-acc6-2e3045407b64"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.846361 4934 generic.go:334] "Generic (PLEG): container finished" podID="a30daf4c-06f2-48f5-b57f-f6704fae13a9" containerID="65dbab7a03568a76645932c0b29b78b90ef3f3460bb8573a3ba9cd75761a9154" exitCode=0 Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.846402 4934 generic.go:334] "Generic (PLEG): container finished" podID="a30daf4c-06f2-48f5-b57f-f6704fae13a9" containerID="ba2dbd7671783785d215106bb6b057dd861995948702cd2be58811e1a9b4c38e" exitCode=0 Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.846452 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-bfc5bc4b9-crmsr" event={"ID":"a30daf4c-06f2-48f5-b57f-f6704fae13a9","Type":"ContainerDied","Data":"65dbab7a03568a76645932c0b29b78b90ef3f3460bb8573a3ba9cd75761a9154"} Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.846485 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-bfc5bc4b9-crmsr" event={"ID":"a30daf4c-06f2-48f5-b57f-f6704fae13a9","Type":"ContainerDied","Data":"ba2dbd7671783785d215106bb6b057dd861995948702cd2be58811e1a9b4c38e"} Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.846500 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-bfc5bc4b9-crmsr" event={"ID":"a30daf4c-06f2-48f5-b57f-f6704fae13a9","Type":"ContainerDied","Data":"83cc72498710eda3ccb7563ee5eb49f8455310af6525b949f517589cb3b7027e"} Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.846513 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83cc72498710eda3ccb7563ee5eb49f8455310af6525b949f517589cb3b7027e" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.866519 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder36fe-account-delete-n2bk5" event={"ID":"bf007c3a-2ba5-4d94-a789-a5413787d6d8","Type":"ContainerStarted","Data":"343e2c45aadc061e539aead3e470e91d2b221db608fa86bf327bb7d852fe8a92"} Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.870014 4934 generic.go:334] "Generic (PLEG): container finished" podID="04aff378-509c-49be-bf4c-2d59111fc910" containerID="0c87adf381aab6824e14a94acbf03d404203f8ec3721590d4efd316ce48e8dd8" exitCode=143 Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.870084 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"04aff378-509c-49be-bf4c-2d59111fc910","Type":"ContainerDied","Data":"0c87adf381aab6824e14a94acbf03d404203f8ec3721590d4efd316ce48e8dd8"} Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.872240 4934 generic.go:334] "Generic (PLEG): container finished" podID="9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f" containerID="b9baba9779ef141c5b9d9cb5a3a64ef9893f0e264bbcdb5320d4acc16116a004" exitCode=0 Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.872302 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f","Type":"ContainerDied","Data":"b9baba9779ef141c5b9d9cb5a3a64ef9893f0e264bbcdb5320d4acc16116a004"} Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.872326 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f","Type":"ContainerDied","Data":"0697d241dfeac3d026f3b9bbabc143e393f12b6a58f2588141d9ee2fd2b2b1f0"} Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.872341 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0697d241dfeac3d026f3b9bbabc143e393f12b6a58f2588141d9ee2fd2b2b1f0" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.873284 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/547c6c6d-0f01-4feb-acc6-2e3045407b64-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "547c6c6d-0f01-4feb-acc6-2e3045407b64" (UID: "547c6c6d-0f01-4feb-acc6-2e3045407b64"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.876094 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance82a2-account-delete-sv9vf" event={"ID":"8679bb63-5c56-42a3-9b2c-42303e3afef6","Type":"ContainerStarted","Data":"d6b82a35f04bd2f56ef75ad50a445e0346ac7b81813f75928def62ce66c8ec56"} Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.881666 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_6061002c-7242-4d02-b868-6838b39f30fa/ovsdbserver-sb/0.log" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.881741 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6061002c-7242-4d02-b868-6838b39f30fa","Type":"ContainerDied","Data":"cd21fda52b3b651db6f76212880c3304577978cd48904a4a334ce44fe9c84cd5"} Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.881826 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.886361 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement5c1f-account-delete-2jjrd" event={"ID":"5ceb70f8-5cc3-4e91-9bc1-6fd0b7575efc","Type":"ContainerStarted","Data":"7ed9155cd1ead1214836ae38ef467cb61192842e0c3254a04b353ba0719115b2"} Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.894182 4934 generic.go:334] "Generic (PLEG): container finished" podID="547c6c6d-0f01-4feb-acc6-2e3045407b64" containerID="c52d38070f0e71890f03b32d417b88c6062d9efcaddea407689d3ce9f0ed9989" exitCode=0 Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.894264 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6bf48f9d67-g95nd" event={"ID":"547c6c6d-0f01-4feb-acc6-2e3045407b64","Type":"ContainerDied","Data":"c52d38070f0e71890f03b32d417b88c6062d9efcaddea407689d3ce9f0ed9989"} Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.894291 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6bf48f9d67-g95nd" event={"ID":"547c6c6d-0f01-4feb-acc6-2e3045407b64","Type":"ContainerDied","Data":"fb60fab0d9fd014001fa386851aec18c4189763f270a448ffbeee76e27776329"} Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.894306 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6bf48f9d67-g95nd" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.898259 4934 generic.go:334] "Generic (PLEG): container finished" podID="125b40d7-516c-432f-ae51-703f2b18068d" containerID="295f3add8c46cf850cd886f2d08991a6be7f2acfd564e68d63ce13cc92ba58df" exitCode=0 Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.898342 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"125b40d7-516c-432f-ae51-703f2b18068d","Type":"ContainerDied","Data":"295f3add8c46cf850cd886f2d08991a6be7f2acfd564e68d63ce13cc92ba58df"} Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.901212 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85b4cf56ff-zr6h2" event={"ID":"20b39d07-1559-4d11-b9ff-6ef97142a58a","Type":"ContainerDied","Data":"a2f8d8a5389fc04137cb0d450fa6ab57ff03fe6a761dfa3a3bad6b4d03aa3c82"} Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.902556 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi26a6-account-delete-ng2mb" event={"ID":"a7f7a4d6-c246-4c88-85e1-573fd07ec33f","Type":"ContainerStarted","Data":"bc902f497027122db034b741d4e334263ead76a6fc25aa6afc21ca8303950f02"} Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.930091 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/547c6c6d-0f01-4feb-acc6-2e3045407b64-config-data" (OuterVolumeSpecName: "config-data") pod "547c6c6d-0f01-4feb-acc6-2e3045407b64" (UID: "547c6c6d-0f01-4feb-acc6-2e3045407b64"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.940882 4934 generic.go:334] "Generic (PLEG): container finished" podID="d15b0450-e100-4ee7-ad62-88601eedc4f2" containerID="2e7e1ea7602c5dbc0c4396593163b8efff5196b8cba864f827def060e37814e7" exitCode=0 Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.940927 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-b97fc8746-vlx9z" event={"ID":"d15b0450-e100-4ee7-ad62-88601eedc4f2","Type":"ContainerDied","Data":"2e7e1ea7602c5dbc0c4396593163b8efff5196b8cba864f827def060e37814e7"} Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.940953 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-b97fc8746-vlx9z" event={"ID":"d15b0450-e100-4ee7-ad62-88601eedc4f2","Type":"ContainerDied","Data":"7eab5abdc57629a153a839021a55ef264408229deaf5d32392f17de6def77d2f"} Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.940988 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7eab5abdc57629a153a839021a55ef264408229deaf5d32392f17de6def77d2f" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.946135 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/547c6c6d-0f01-4feb-acc6-2e3045407b64-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:13 crc kubenswrapper[4934]: E1002 10:11:13.946219 4934 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.946236 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxzdw\" (UniqueName: \"kubernetes.io/projected/547c6c6d-0f01-4feb-acc6-2e3045407b64-kube-api-access-vxzdw\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:13 crc kubenswrapper[4934]: E1002 10:11:13.946283 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/97560469-c459-48f4-901e-54eaded32537-config-data podName:97560469-c459-48f4-901e-54eaded32537 nodeName:}" failed. No retries permitted until 2025-10-02 10:11:17.94626296 +0000 UTC m=+1349.698904482 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/97560469-c459-48f4-901e-54eaded32537-config-data") pod "rabbitmq-server-0" (UID: "97560469-c459-48f4-901e-54eaded32537") : configmap "rabbitmq-config-data" not found Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.947270 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/547c6c6d-0f01-4feb-acc6-2e3045407b64-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.947302 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/547c6c6d-0f01-4feb-acc6-2e3045407b64-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.947317 4934 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/547c6c6d-0f01-4feb-acc6-2e3045407b64-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.950400 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.950720 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4becb20a-8bf6-458d-856e-c8d6cd43f797" containerName="ceilometer-central-agent" containerID="cri-o://bf2e8afd472c84d3385087f6698782dc92f0a62388c560ef85d4823b8156cfb7" gracePeriod=30 Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.954774 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4becb20a-8bf6-458d-856e-c8d6cd43f797" containerName="proxy-httpd" containerID="cri-o://8cdeead710c71e753ff7516bcb319070ccf923a448f85b9b738b5bbc64e5abd1" gracePeriod=30 Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.954797 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4becb20a-8bf6-458d-856e-c8d6cd43f797" containerName="sg-core" containerID="cri-o://1b462ec2465f66c7ac2fdf185d407ece6092286b180138bdcd246fc6359029b3" gracePeriod=30 Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.954900 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4becb20a-8bf6-458d-856e-c8d6cd43f797" containerName="ceilometer-notification-agent" containerID="cri-o://8e5fcc5a4c69ba03bdc17a733d109d36416615950178c66556ce847df4a0f0cf" gracePeriod=30 Oct 02 10:11:13 crc kubenswrapper[4934]: E1002 10:11:13.959916 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b7f8b726c37d4fd8d67700b1306d05906985514bcb56092fd5f3e2779cf2dc8a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 10:11:13 crc kubenswrapper[4934]: E1002 10:11:13.980512 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b7f8b726c37d4fd8d67700b1306d05906985514bcb56092fd5f3e2779cf2dc8a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.999446 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 10:11:13 crc kubenswrapper[4934]: I1002 10:11:13.999736 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="958bc92f-fd6d-4ba0-a21f-8f1c58bac789" containerName="kube-state-metrics" containerID="cri-o://877a883b01af6709bd26ee32e0a326b213d51f06b683c07ca0c300ae16ad52db" gracePeriod=30 Oct 02 10:11:14 crc kubenswrapper[4934]: E1002 10:11:14.024637 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b7f8b726c37d4fd8d67700b1306d05906985514bcb56092fd5f3e2779cf2dc8a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 10:11:14 crc kubenswrapper[4934]: E1002 10:11:14.024757 4934 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="06adab43-0177-440c-aeef-9386345160c4" containerName="nova-scheduler-scheduler" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.101964 4934 scope.go:117] "RemoveContainer" containerID="aac633fecdf45004f95036aeac08948cfcfe75a04ef1942aef60308cc2356517" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.136219 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.136429 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="f43ed4d4-3637-4ef6-a914-3a751cba8b4e" containerName="memcached" containerID="cri-o://9d4708b7023ab963058c9d146e0e8bf04766c8e6b064f4e76425c8490490d0bd" gracePeriod=30 Oct 02 10:11:14 crc kubenswrapper[4934]: E1002 10:11:14.143001 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 38289b907213ccb98a8b74c9827dd591276a04d2139306b7c6ab58968debd430 is running failed: container process not found" containerID="38289b907213ccb98a8b74c9827dd591276a04d2139306b7c6ab58968debd430" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 02 10:11:14 crc kubenswrapper[4934]: E1002 10:11:14.143396 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 38289b907213ccb98a8b74c9827dd591276a04d2139306b7c6ab58968debd430 is running failed: container process not found" containerID="38289b907213ccb98a8b74c9827dd591276a04d2139306b7c6ab58968debd430" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 02 10:11:14 crc kubenswrapper[4934]: E1002 10:11:14.143698 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 38289b907213ccb98a8b74c9827dd591276a04d2139306b7c6ab58968debd430 is running failed: container process not found" containerID="38289b907213ccb98a8b74c9827dd591276a04d2139306b7c6ab58968debd430" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 02 10:11:14 crc kubenswrapper[4934]: E1002 10:11:14.143735 4934 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 38289b907213ccb98a8b74c9827dd591276a04d2139306b7c6ab58968debd430 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="49a440a9-1a3f-4ee9-bf28-7996fe6f222d" containerName="nova-cell1-conductor-conductor" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.173657 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-gwpkm"] Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.191753 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-vzw77"] Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.202721 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-b97fc8746-vlx9z" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.258890 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtknw\" (UniqueName: \"kubernetes.io/projected/d15b0450-e100-4ee7-ad62-88601eedc4f2-kube-api-access-vtknw\") pod \"d15b0450-e100-4ee7-ad62-88601eedc4f2\" (UID: \"d15b0450-e100-4ee7-ad62-88601eedc4f2\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.258993 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d15b0450-e100-4ee7-ad62-88601eedc4f2-config-data\") pod \"d15b0450-e100-4ee7-ad62-88601eedc4f2\" (UID: \"d15b0450-e100-4ee7-ad62-88601eedc4f2\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.259027 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d15b0450-e100-4ee7-ad62-88601eedc4f2-logs\") pod \"d15b0450-e100-4ee7-ad62-88601eedc4f2\" (UID: \"d15b0450-e100-4ee7-ad62-88601eedc4f2\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.259055 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d15b0450-e100-4ee7-ad62-88601eedc4f2-combined-ca-bundle\") pod \"d15b0450-e100-4ee7-ad62-88601eedc4f2\" (UID: \"d15b0450-e100-4ee7-ad62-88601eedc4f2\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.259076 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d15b0450-e100-4ee7-ad62-88601eedc4f2-config-data-custom\") pod \"d15b0450-e100-4ee7-ad62-88601eedc4f2\" (UID: \"d15b0450-e100-4ee7-ad62-88601eedc4f2\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.259872 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d15b0450-e100-4ee7-ad62-88601eedc4f2-logs" (OuterVolumeSpecName: "logs") pod "d15b0450-e100-4ee7-ad62-88601eedc4f2" (UID: "d15b0450-e100-4ee7-ad62-88601eedc4f2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.268399 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.270402 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-gwpkm"] Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.272321 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d15b0450-e100-4ee7-ad62-88601eedc4f2-kube-api-access-vtknw" (OuterVolumeSpecName: "kube-api-access-vtknw") pod "d15b0450-e100-4ee7-ad62-88601eedc4f2" (UID: "d15b0450-e100-4ee7-ad62-88601eedc4f2"). InnerVolumeSpecName "kube-api-access-vtknw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.277437 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.279812 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d15b0450-e100-4ee7-ad62-88601eedc4f2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d15b0450-e100-4ee7-ad62-88601eedc4f2" (UID: "d15b0450-e100-4ee7-ad62-88601eedc4f2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.289470 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-vzw77"] Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.291562 4934 scope.go:117] "RemoveContainer" containerID="6ae422d564bec96f2ecabdd62c9fb7ad073e005254728994f68b5086ec8bc468" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.301078 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.305304 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-c25xv"] Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.312470 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-c25xv"] Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.320563 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.345565 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-7b7b4b7b85-9llg8"] Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.345860 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-7b7b4b7b85-9llg8" podUID="02feb54e-6b14-42ec-a22d-524a6005e2fd" containerName="keystone-api" containerID="cri-o://e7f1b0e43bf1c9bf7da08e7461ca997f5bf6711c2a32f9d4be458d1aff9e7cdc" gracePeriod=30 Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.356766 4934 scope.go:117] "RemoveContainer" containerID="fd9f74a1d204e8b395d9002c1eb07158525244b64cff40fb8c86051204dcedaa" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.365025 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d15b0450-e100-4ee7-ad62-88601eedc4f2-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.365056 4934 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d15b0450-e100-4ee7-ad62-88601eedc4f2-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.365067 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtknw\" (UniqueName: \"kubernetes.io/projected/d15b0450-e100-4ee7-ad62-88601eedc4f2-kube-api-access-vtknw\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.389880 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.409769 4934 scope.go:117] "RemoveContainer" containerID="abc6a88d34ca1719e3b11b44524da713532a93344a46bb1626798816c0caa7a2" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.409983 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-mrnhj"] Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.424363 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-mrnhj"] Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.434211 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-b73e-account-create-mnzcm"] Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.455200 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-b73e-account-create-mnzcm"] Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.465860 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49a440a9-1a3f-4ee9-bf28-7996fe6f222d-combined-ca-bundle\") pod \"49a440a9-1a3f-4ee9-bf28-7996fe6f222d\" (UID: \"49a440a9-1a3f-4ee9-bf28-7996fe6f222d\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.465926 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzczs\" (UniqueName: \"kubernetes.io/projected/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-kube-api-access-mzczs\") pod \"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f\" (UID: \"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.465980 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a30daf4c-06f2-48f5-b57f-f6704fae13a9-run-httpd\") pod \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.466018 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-nova-novncproxy-tls-certs\") pod \"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f\" (UID: \"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.466044 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcxvb\" (UniqueName: \"kubernetes.io/projected/a30daf4c-06f2-48f5-b57f-f6704fae13a9-kube-api-access-xcxvb\") pod \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.466071 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/125b40d7-516c-432f-ae51-703f2b18068d-galera-tls-certs\") pod \"125b40d7-516c-432f-ae51-703f2b18068d\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.466130 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"125b40d7-516c-432f-ae51-703f2b18068d\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.466169 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/125b40d7-516c-432f-ae51-703f2b18068d-operator-scripts\") pod \"125b40d7-516c-432f-ae51-703f2b18068d\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.466196 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a30daf4c-06f2-48f5-b57f-f6704fae13a9-log-httpd\") pod \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.466236 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgqm8\" (UniqueName: \"kubernetes.io/projected/125b40d7-516c-432f-ae51-703f2b18068d-kube-api-access-bgqm8\") pod \"125b40d7-516c-432f-ae51-703f2b18068d\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.466263 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/125b40d7-516c-432f-ae51-703f2b18068d-config-data-generated\") pod \"125b40d7-516c-432f-ae51-703f2b18068d\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.466292 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-combined-ca-bundle\") pod \"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f\" (UID: \"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.466700 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a30daf4c-06f2-48f5-b57f-f6704fae13a9-internal-tls-certs\") pod \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.466743 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gd8fm\" (UniqueName: \"kubernetes.io/projected/49a440a9-1a3f-4ee9-bf28-7996fe6f222d-kube-api-access-gd8fm\") pod \"49a440a9-1a3f-4ee9-bf28-7996fe6f222d\" (UID: \"49a440a9-1a3f-4ee9-bf28-7996fe6f222d\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.466792 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49a440a9-1a3f-4ee9-bf28-7996fe6f222d-config-data\") pod \"49a440a9-1a3f-4ee9-bf28-7996fe6f222d\" (UID: \"49a440a9-1a3f-4ee9-bf28-7996fe6f222d\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.466827 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/125b40d7-516c-432f-ae51-703f2b18068d-kolla-config\") pod \"125b40d7-516c-432f-ae51-703f2b18068d\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.466851 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a30daf4c-06f2-48f5-b57f-f6704fae13a9-config-data\") pod \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.466874 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a30daf4c-06f2-48f5-b57f-f6704fae13a9-combined-ca-bundle\") pod \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.466912 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-config-data\") pod \"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f\" (UID: \"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.466956 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a30daf4c-06f2-48f5-b57f-f6704fae13a9-etc-swift\") pod \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.466978 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/125b40d7-516c-432f-ae51-703f2b18068d-config-data-default\") pod \"125b40d7-516c-432f-ae51-703f2b18068d\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.467012 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-vencrypt-tls-certs\") pod \"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f\" (UID: \"9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.467040 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/125b40d7-516c-432f-ae51-703f2b18068d-combined-ca-bundle\") pod \"125b40d7-516c-432f-ae51-703f2b18068d\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.467060 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a30daf4c-06f2-48f5-b57f-f6704fae13a9-public-tls-certs\") pod \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\" (UID: \"a30daf4c-06f2-48f5-b57f-f6704fae13a9\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.467122 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/125b40d7-516c-432f-ae51-703f2b18068d-secrets\") pod \"125b40d7-516c-432f-ae51-703f2b18068d\" (UID: \"125b40d7-516c-432f-ae51-703f2b18068d\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.469138 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a30daf4c-06f2-48f5-b57f-f6704fae13a9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a30daf4c-06f2-48f5-b57f-f6704fae13a9" (UID: "a30daf4c-06f2-48f5-b57f-f6704fae13a9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.472406 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/125b40d7-516c-432f-ae51-703f2b18068d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "125b40d7-516c-432f-ae51-703f2b18068d" (UID: "125b40d7-516c-432f-ae51-703f2b18068d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.472527 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/125b40d7-516c-432f-ae51-703f2b18068d-secrets" (OuterVolumeSpecName: "secrets") pod "125b40d7-516c-432f-ae51-703f2b18068d" (UID: "125b40d7-516c-432f-ae51-703f2b18068d"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.472750 4934 scope.go:117] "RemoveContainer" containerID="c52d38070f0e71890f03b32d417b88c6062d9efcaddea407689d3ce9f0ed9989" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.473051 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/125b40d7-516c-432f-ae51-703f2b18068d-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "125b40d7-516c-432f-ae51-703f2b18068d" (UID: "125b40d7-516c-432f-ae51-703f2b18068d"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.473657 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a30daf4c-06f2-48f5-b57f-f6704fae13a9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a30daf4c-06f2-48f5-b57f-f6704fae13a9" (UID: "a30daf4c-06f2-48f5-b57f-f6704fae13a9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.474150 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/125b40d7-516c-432f-ae51-703f2b18068d-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "125b40d7-516c-432f-ae51-703f2b18068d" (UID: "125b40d7-516c-432f-ae51-703f2b18068d"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.475079 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/125b40d7-516c-432f-ae51-703f2b18068d-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "125b40d7-516c-432f-ae51-703f2b18068d" (UID: "125b40d7-516c-432f-ae51-703f2b18068d"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.480025 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.491337 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.495889 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-kube-api-access-mzczs" (OuterVolumeSpecName: "kube-api-access-mzczs") pod "9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f" (UID: "9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f"). InnerVolumeSpecName "kube-api-access-mzczs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.496117 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49a440a9-1a3f-4ee9-bf28-7996fe6f222d-kube-api-access-gd8fm" (OuterVolumeSpecName: "kube-api-access-gd8fm") pod "49a440a9-1a3f-4ee9-bf28-7996fe6f222d" (UID: "49a440a9-1a3f-4ee9-bf28-7996fe6f222d"). InnerVolumeSpecName "kube-api-access-gd8fm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.496245 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d15b0450-e100-4ee7-ad62-88601eedc4f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d15b0450-e100-4ee7-ad62-88601eedc4f2" (UID: "d15b0450-e100-4ee7-ad62-88601eedc4f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.496679 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a30daf4c-06f2-48f5-b57f-f6704fae13a9-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "a30daf4c-06f2-48f5-b57f-f6704fae13a9" (UID: "a30daf4c-06f2-48f5-b57f-f6704fae13a9"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.503407 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.505025 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/125b40d7-516c-432f-ae51-703f2b18068d-kube-api-access-bgqm8" (OuterVolumeSpecName: "kube-api-access-bgqm8") pod "125b40d7-516c-432f-ae51-703f2b18068d" (UID: "125b40d7-516c-432f-ae51-703f2b18068d"). InnerVolumeSpecName "kube-api-access-bgqm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.509419 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a30daf4c-06f2-48f5-b57f-f6704fae13a9-kube-api-access-xcxvb" (OuterVolumeSpecName: "kube-api-access-xcxvb") pod "a30daf4c-06f2-48f5-b57f-f6704fae13a9" (UID: "a30daf4c-06f2-48f5-b57f-f6704fae13a9"). InnerVolumeSpecName "kube-api-access-xcxvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.510389 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.518817 4934 scope.go:117] "RemoveContainer" containerID="ca5c961136b99e80a8ae0bb9c203d870c7355071cc4a476bdb7c41606bdef8b7" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.521254 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-l476l"] Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.526791 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-l476l"] Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.542076 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85b4cf56ff-zr6h2"] Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.542567 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85b4cf56ff-zr6h2"] Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.543672 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-6bf48f9d67-g95nd"] Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.544791 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "mysql-db") pod "125b40d7-516c-432f-ae51-703f2b18068d" (UID: "125b40d7-516c-432f-ae51-703f2b18068d"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.570098 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-6bf48f9d67-g95nd"] Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.571287 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcxvb\" (UniqueName: \"kubernetes.io/projected/a30daf4c-06f2-48f5-b57f-f6704fae13a9-kube-api-access-xcxvb\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.571314 4934 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.571324 4934 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/125b40d7-516c-432f-ae51-703f2b18068d-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.571333 4934 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a30daf4c-06f2-48f5-b57f-f6704fae13a9-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.571341 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgqm8\" (UniqueName: \"kubernetes.io/projected/125b40d7-516c-432f-ae51-703f2b18068d-kube-api-access-bgqm8\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.571351 4934 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/125b40d7-516c-432f-ae51-703f2b18068d-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.571359 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d15b0450-e100-4ee7-ad62-88601eedc4f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.571367 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gd8fm\" (UniqueName: \"kubernetes.io/projected/49a440a9-1a3f-4ee9-bf28-7996fe6f222d-kube-api-access-gd8fm\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.571375 4934 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/125b40d7-516c-432f-ae51-703f2b18068d-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.571383 4934 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a30daf4c-06f2-48f5-b57f-f6704fae13a9-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.571391 4934 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/125b40d7-516c-432f-ae51-703f2b18068d-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.571398 4934 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/125b40d7-516c-432f-ae51-703f2b18068d-secrets\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.571406 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzczs\" (UniqueName: \"kubernetes.io/projected/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-kube-api-access-mzczs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.571413 4934 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a30daf4c-06f2-48f5-b57f-f6704fae13a9-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.572337 4934 scope.go:117] "RemoveContainer" containerID="c52d38070f0e71890f03b32d417b88c6062d9efcaddea407689d3ce9f0ed9989" Oct 02 10:11:14 crc kubenswrapper[4934]: E1002 10:11:14.572830 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c52d38070f0e71890f03b32d417b88c6062d9efcaddea407689d3ce9f0ed9989\": container with ID starting with c52d38070f0e71890f03b32d417b88c6062d9efcaddea407689d3ce9f0ed9989 not found: ID does not exist" containerID="c52d38070f0e71890f03b32d417b88c6062d9efcaddea407689d3ce9f0ed9989" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.572857 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c52d38070f0e71890f03b32d417b88c6062d9efcaddea407689d3ce9f0ed9989"} err="failed to get container status \"c52d38070f0e71890f03b32d417b88c6062d9efcaddea407689d3ce9f0ed9989\": rpc error: code = NotFound desc = could not find container \"c52d38070f0e71890f03b32d417b88c6062d9efcaddea407689d3ce9f0ed9989\": container with ID starting with c52d38070f0e71890f03b32d417b88c6062d9efcaddea407689d3ce9f0ed9989 not found: ID does not exist" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.572878 4934 scope.go:117] "RemoveContainer" containerID="ca5c961136b99e80a8ae0bb9c203d870c7355071cc4a476bdb7c41606bdef8b7" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.596672 4934 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 02 10:11:14 crc kubenswrapper[4934]: E1002 10:11:14.606889 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca5c961136b99e80a8ae0bb9c203d870c7355071cc4a476bdb7c41606bdef8b7\": container with ID starting with ca5c961136b99e80a8ae0bb9c203d870c7355071cc4a476bdb7c41606bdef8b7 not found: ID does not exist" containerID="ca5c961136b99e80a8ae0bb9c203d870c7355071cc4a476bdb7c41606bdef8b7" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.606945 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca5c961136b99e80a8ae0bb9c203d870c7355071cc4a476bdb7c41606bdef8b7"} err="failed to get container status \"ca5c961136b99e80a8ae0bb9c203d870c7355071cc4a476bdb7c41606bdef8b7\": rpc error: code = NotFound desc = could not find container \"ca5c961136b99e80a8ae0bb9c203d870c7355071cc4a476bdb7c41606bdef8b7\": container with ID starting with ca5c961136b99e80a8ae0bb9c203d870c7355071cc4a476bdb7c41606bdef8b7 not found: ID does not exist" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.606977 4934 scope.go:117] "RemoveContainer" containerID="aac633fecdf45004f95036aeac08948cfcfe75a04ef1942aef60308cc2356517" Oct 02 10:11:14 crc kubenswrapper[4934]: E1002 10:11:14.610464 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aac633fecdf45004f95036aeac08948cfcfe75a04ef1942aef60308cc2356517\": container with ID starting with aac633fecdf45004f95036aeac08948cfcfe75a04ef1942aef60308cc2356517 not found: ID does not exist" containerID="aac633fecdf45004f95036aeac08948cfcfe75a04ef1942aef60308cc2356517" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.610508 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aac633fecdf45004f95036aeac08948cfcfe75a04ef1942aef60308cc2356517"} err="failed to get container status \"aac633fecdf45004f95036aeac08948cfcfe75a04ef1942aef60308cc2356517\": rpc error: code = NotFound desc = could not find container \"aac633fecdf45004f95036aeac08948cfcfe75a04ef1942aef60308cc2356517\": container with ID starting with aac633fecdf45004f95036aeac08948cfcfe75a04ef1942aef60308cc2356517 not found: ID does not exist" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.610534 4934 scope.go:117] "RemoveContainer" containerID="6ae422d564bec96f2ecabdd62c9fb7ad073e005254728994f68b5086ec8bc468" Oct 02 10:11:14 crc kubenswrapper[4934]: E1002 10:11:14.611004 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ae422d564bec96f2ecabdd62c9fb7ad073e005254728994f68b5086ec8bc468\": container with ID starting with 6ae422d564bec96f2ecabdd62c9fb7ad073e005254728994f68b5086ec8bc468 not found: ID does not exist" containerID="6ae422d564bec96f2ecabdd62c9fb7ad073e005254728994f68b5086ec8bc468" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.611037 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ae422d564bec96f2ecabdd62c9fb7ad073e005254728994f68b5086ec8bc468"} err="failed to get container status \"6ae422d564bec96f2ecabdd62c9fb7ad073e005254728994f68b5086ec8bc468\": rpc error: code = NotFound desc = could not find container \"6ae422d564bec96f2ecabdd62c9fb7ad073e005254728994f68b5086ec8bc468\": container with ID starting with 6ae422d564bec96f2ecabdd62c9fb7ad073e005254728994f68b5086ec8bc468 not found: ID does not exist" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.615902 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d15b0450-e100-4ee7-ad62-88601eedc4f2-config-data" (OuterVolumeSpecName: "config-data") pod "d15b0450-e100-4ee7-ad62-88601eedc4f2" (UID: "d15b0450-e100-4ee7-ad62-88601eedc4f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.628764 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f" (UID: "9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.634003 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49a440a9-1a3f-4ee9-bf28-7996fe6f222d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "49a440a9-1a3f-4ee9-bf28-7996fe6f222d" (UID: "49a440a9-1a3f-4ee9-bf28-7996fe6f222d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.634285 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49a440a9-1a3f-4ee9-bf28-7996fe6f222d-config-data" (OuterVolumeSpecName: "config-data") pod "49a440a9-1a3f-4ee9-bf28-7996fe6f222d" (UID: "49a440a9-1a3f-4ee9-bf28-7996fe6f222d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.657752 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/125b40d7-516c-432f-ae51-703f2b18068d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "125b40d7-516c-432f-ae51-703f2b18068d" (UID: "125b40d7-516c-432f-ae51-703f2b18068d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.660992 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-config-data" (OuterVolumeSpecName: "config-data") pod "9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f" (UID: "9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.661196 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.667726 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a30daf4c-06f2-48f5-b57f-f6704fae13a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a30daf4c-06f2-48f5-b57f-f6704fae13a9" (UID: "a30daf4c-06f2-48f5-b57f-f6704fae13a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.670052 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f" (UID: "9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.670736 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a30daf4c-06f2-48f5-b57f-f6704fae13a9-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a30daf4c-06f2-48f5-b57f-f6704fae13a9" (UID: "a30daf4c-06f2-48f5-b57f-f6704fae13a9"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.673537 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49a440a9-1a3f-4ee9-bf28-7996fe6f222d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.674403 4934 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.674476 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d15b0450-e100-4ee7-ad62-88601eedc4f2-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.674530 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.674594 4934 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a30daf4c-06f2-48f5-b57f-f6704fae13a9-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.674664 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49a440a9-1a3f-4ee9-bf28-7996fe6f222d-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.674716 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a30daf4c-06f2-48f5-b57f-f6704fae13a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.674765 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.674815 4934 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.674866 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/125b40d7-516c-432f-ae51-703f2b18068d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.683540 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a30daf4c-06f2-48f5-b57f-f6704fae13a9-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a30daf4c-06f2-48f5-b57f-f6704fae13a9" (UID: "a30daf4c-06f2-48f5-b57f-f6704fae13a9"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.684664 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/125b40d7-516c-432f-ae51-703f2b18068d-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "125b40d7-516c-432f-ae51-703f2b18068d" (UID: "125b40d7-516c-432f-ae51-703f2b18068d"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.693093 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f" (UID: "9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.708041 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a30daf4c-06f2-48f5-b57f-f6704fae13a9-config-data" (OuterVolumeSpecName: "config-data") pod "a30daf4c-06f2-48f5-b57f-f6704fae13a9" (UID: "a30daf4c-06f2-48f5-b57f-f6704fae13a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.716464 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="8c9f23b6-92f8-4d0a-8595-84c142adc3dd" containerName="galera" containerID="cri-o://6ed60ff6b58b2bf5ef3faeb5193e2b929c4e3c03eaae93658e8fc29218880ae3" gracePeriod=30 Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.745293 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cgkdl" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.745372 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cgkdl" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.776070 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/958bc92f-fd6d-4ba0-a21f-8f1c58bac789-kube-state-metrics-tls-config\") pod \"958bc92f-fd6d-4ba0-a21f-8f1c58bac789\" (UID: \"958bc92f-fd6d-4ba0-a21f-8f1c58bac789\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.776184 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/958bc92f-fd6d-4ba0-a21f-8f1c58bac789-combined-ca-bundle\") pod \"958bc92f-fd6d-4ba0-a21f-8f1c58bac789\" (UID: \"958bc92f-fd6d-4ba0-a21f-8f1c58bac789\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.776224 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/958bc92f-fd6d-4ba0-a21f-8f1c58bac789-kube-state-metrics-tls-certs\") pod \"958bc92f-fd6d-4ba0-a21f-8f1c58bac789\" (UID: \"958bc92f-fd6d-4ba0-a21f-8f1c58bac789\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.776377 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-np29c\" (UniqueName: \"kubernetes.io/projected/958bc92f-fd6d-4ba0-a21f-8f1c58bac789-kube-api-access-np29c\") pod \"958bc92f-fd6d-4ba0-a21f-8f1c58bac789\" (UID: \"958bc92f-fd6d-4ba0-a21f-8f1c58bac789\") " Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.780347 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a30daf4c-06f2-48f5-b57f-f6704fae13a9-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.780467 4934 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a30daf4c-06f2-48f5-b57f-f6704fae13a9-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.780523 4934 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.780605 4934 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/125b40d7-516c-432f-ae51-703f2b18068d-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.782440 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/958bc92f-fd6d-4ba0-a21f-8f1c58bac789-kube-api-access-np29c" (OuterVolumeSpecName: "kube-api-access-np29c") pod "958bc92f-fd6d-4ba0-a21f-8f1c58bac789" (UID: "958bc92f-fd6d-4ba0-a21f-8f1c58bac789"). InnerVolumeSpecName "kube-api-access-np29c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.786355 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="cd41dc17-6659-4b43-9758-f67d831993ba" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.158:8776/healthcheck\": read tcp 10.217.0.2:51748->10.217.0.158:8776: read: connection reset by peer" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.809928 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cgkdl" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.815933 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/958bc92f-fd6d-4ba0-a21f-8f1c58bac789-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "958bc92f-fd6d-4ba0-a21f-8f1c58bac789" (UID: "958bc92f-fd6d-4ba0-a21f-8f1c58bac789"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.816311 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/958bc92f-fd6d-4ba0-a21f-8f1c58bac789-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "958bc92f-fd6d-4ba0-a21f-8f1c58bac789" (UID: "958bc92f-fd6d-4ba0-a21f-8f1c58bac789"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.859780 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/958bc92f-fd6d-4ba0-a21f-8f1c58bac789-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "958bc92f-fd6d-4ba0-a21f-8f1c58bac789" (UID: "958bc92f-fd6d-4ba0-a21f-8f1c58bac789"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.888859 4934 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/958bc92f-fd6d-4ba0-a21f-8f1c58bac789-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.888889 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-np29c\" (UniqueName: \"kubernetes.io/projected/958bc92f-fd6d-4ba0-a21f-8f1c58bac789-kube-api-access-np29c\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.888898 4934 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/958bc92f-fd6d-4ba0-a21f-8f1c58bac789-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.888907 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/958bc92f-fd6d-4ba0-a21f-8f1c58bac789-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.927047 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b5b09b9-1368-4f7d-9ed2-0ba034890110" path="/var/lib/kubelet/pods/1b5b09b9-1368-4f7d-9ed2-0ba034890110/volumes" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.927940 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b39d07-1559-4d11-b9ff-6ef97142a58a" path="/var/lib/kubelet/pods/20b39d07-1559-4d11-b9ff-6ef97142a58a/volumes" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.928485 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c705cde-0601-4f0e-97a3-590adbf0da4c" path="/var/lib/kubelet/pods/4c705cde-0601-4f0e-97a3-590adbf0da4c/volumes" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.929482 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="547c6c6d-0f01-4feb-acc6-2e3045407b64" path="/var/lib/kubelet/pods/547c6c6d-0f01-4feb-acc6-2e3045407b64/volumes" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.930221 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6061002c-7242-4d02-b868-6838b39f30fa" path="/var/lib/kubelet/pods/6061002c-7242-4d02-b868-6838b39f30fa/volumes" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.930876 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f4da5aa-1b82-4286-8a3b-68ca8f486c78" path="/var/lib/kubelet/pods/6f4da5aa-1b82-4286-8a3b-68ca8f486c78/volumes" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.931921 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7214291c-3c8e-4173-a6fd-ef16170258e2" path="/var/lib/kubelet/pods/7214291c-3c8e-4173-a6fd-ef16170258e2/volumes" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.932639 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c99e264-05b3-4cd7-91ae-cabb6d10466a" path="/var/lib/kubelet/pods/9c99e264-05b3-4cd7-91ae-cabb6d10466a/volumes" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.933231 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d816754-2d4c-40e6-acee-218947ab9e19" path="/var/lib/kubelet/pods/9d816754-2d4c-40e6-acee-218947ab9e19/volumes" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.934198 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e14738be-f83f-47ce-b0d9-ceb23d504b46" path="/var/lib/kubelet/pods/e14738be-f83f-47ce-b0d9-ceb23d504b46/volumes" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.934696 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7fcfea2-ee1e-4ef2-ac7b-2948b0f289ac" path="/var/lib/kubelet/pods/f7fcfea2-ee1e-4ef2-ac7b-2948b0f289ac/volumes" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.954053 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/placement-649ccd9666-dh5rs" podUID="ecc76715-8b05-4529-bd7b-289d7f32eff5" containerName="placement-log" probeResult="failure" output="Get \"https://10.217.0.147:8778/\": read tcp 10.217.0.2:51878->10.217.0.147:8778: read: connection reset by peer" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.954342 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/placement-649ccd9666-dh5rs" podUID="ecc76715-8b05-4529-bd7b-289d7f32eff5" containerName="placement-api" probeResult="failure" output="Get \"https://10.217.0.147:8778/\": read tcp 10.217.0.2:51894->10.217.0.147:8778: read: connection reset by peer" Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.998001 4934 generic.go:334] "Generic (PLEG): container finished" podID="84d2000a-d99f-4320-b178-874940ab7e9d" containerID="391ab06a0bf5924bac18ac518442b3a5539dd7de7e5001512639e7be11a54744" exitCode=0 Oct 02 10:11:14 crc kubenswrapper[4934]: I1002 10:11:14.998056 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"84d2000a-d99f-4320-b178-874940ab7e9d","Type":"ContainerDied","Data":"391ab06a0bf5924bac18ac518442b3a5539dd7de7e5001512639e7be11a54744"} Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.000390 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance82a2-account-delete-sv9vf" event={"ID":"8679bb63-5c56-42a3-9b2c-42303e3afef6","Type":"ContainerStarted","Data":"2696284bd53dcc40f609d9c2ea7a8c80534e0456354a1ab9817e35d0d0a4d9a4"} Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.000510 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance82a2-account-delete-sv9vf" podUID="8679bb63-5c56-42a3-9b2c-42303e3afef6" containerName="mariadb-account-delete" containerID="cri-o://2696284bd53dcc40f609d9c2ea7a8c80534e0456354a1ab9817e35d0d0a4d9a4" gracePeriod=30 Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.007515 4934 generic.go:334] "Generic (PLEG): container finished" podID="958bc92f-fd6d-4ba0-a21f-8f1c58bac789" containerID="877a883b01af6709bd26ee32e0a326b213d51f06b683c07ca0c300ae16ad52db" exitCode=2 Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.007598 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"958bc92f-fd6d-4ba0-a21f-8f1c58bac789","Type":"ContainerDied","Data":"877a883b01af6709bd26ee32e0a326b213d51f06b683c07ca0c300ae16ad52db"} Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.007630 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"958bc92f-fd6d-4ba0-a21f-8f1c58bac789","Type":"ContainerDied","Data":"6653b74890db6090b03cbdefd5f76446661ab9976bd658ec66207c6f4414e1e8"} Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.007652 4934 scope.go:117] "RemoveContainer" containerID="877a883b01af6709bd26ee32e0a326b213d51f06b683c07ca0c300ae16ad52db" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.007787 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.019117 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"125b40d7-516c-432f-ae51-703f2b18068d","Type":"ContainerDied","Data":"88252b1154831e865b61e0dd5b582c526915e28c56e8dd89163117a6b8f226a8"} Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.019230 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.041778 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement5c1f-account-delete-2jjrd" event={"ID":"5ceb70f8-5cc3-4e91-9bc1-6fd0b7575efc","Type":"ContainerStarted","Data":"3bc9955bd39440cdf80a78378bd0626b6c3f71633a26259bbf572670b66ff302"} Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.041935 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement5c1f-account-delete-2jjrd" podUID="5ceb70f8-5cc3-4e91-9bc1-6fd0b7575efc" containerName="mariadb-account-delete" containerID="cri-o://3bc9955bd39440cdf80a78378bd0626b6c3f71633a26259bbf572670b66ff302" gracePeriod=30 Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.055025 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder36fe-account-delete-n2bk5" event={"ID":"bf007c3a-2ba5-4d94-a789-a5413787d6d8","Type":"ContainerStarted","Data":"114be7966ff7fb8bfc8736837ac939a0abe593bac1ebf6b0afb5973b3356ee6d"} Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.055271 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder36fe-account-delete-n2bk5" podUID="bf007c3a-2ba5-4d94-a789-a5413787d6d8" containerName="mariadb-account-delete" containerID="cri-o://114be7966ff7fb8bfc8736837ac939a0abe593bac1ebf6b0afb5973b3356ee6d" gracePeriod=30 Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.078135 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance82a2-account-delete-sv9vf" podStartSLOduration=6.078112976 podStartE2EDuration="6.078112976s" podCreationTimestamp="2025-10-02 10:11:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:11:15.014685946 +0000 UTC m=+1346.767327468" watchObservedRunningTime="2025-10-02 10:11:15.078112976 +0000 UTC m=+1346.830754498" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.080198 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement5c1f-account-delete-2jjrd" podStartSLOduration=6.080188744 podStartE2EDuration="6.080188744s" podCreationTimestamp="2025-10-02 10:11:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:11:15.076398479 +0000 UTC m=+1346.829040001" watchObservedRunningTime="2025-10-02 10:11:15.080188744 +0000 UTC m=+1346.832830266" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.085630 4934 generic.go:334] "Generic (PLEG): container finished" podID="4becb20a-8bf6-458d-856e-c8d6cd43f797" containerID="8cdeead710c71e753ff7516bcb319070ccf923a448f85b9b738b5bbc64e5abd1" exitCode=0 Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.085666 4934 generic.go:334] "Generic (PLEG): container finished" podID="4becb20a-8bf6-458d-856e-c8d6cd43f797" containerID="1b462ec2465f66c7ac2fdf185d407ece6092286b180138bdcd246fc6359029b3" exitCode=2 Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.085674 4934 generic.go:334] "Generic (PLEG): container finished" podID="4becb20a-8bf6-458d-856e-c8d6cd43f797" containerID="bf2e8afd472c84d3385087f6698782dc92f0a62388c560ef85d4823b8156cfb7" exitCode=0 Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.085765 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4becb20a-8bf6-458d-856e-c8d6cd43f797","Type":"ContainerDied","Data":"8cdeead710c71e753ff7516bcb319070ccf923a448f85b9b738b5bbc64e5abd1"} Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.085796 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4becb20a-8bf6-458d-856e-c8d6cd43f797","Type":"ContainerDied","Data":"1b462ec2465f66c7ac2fdf185d407ece6092286b180138bdcd246fc6359029b3"} Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.085808 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4becb20a-8bf6-458d-856e-c8d6cd43f797","Type":"ContainerDied","Data":"bf2e8afd472c84d3385087f6698782dc92f0a62388c560ef85d4823b8156cfb7"} Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.096853 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi26a6-account-delete-ng2mb" event={"ID":"a7f7a4d6-c246-4c88-85e1-573fd07ec33f","Type":"ContainerStarted","Data":"4251b40d65b71246a67772b89828963553528ad5d1e49e7473e8cac466296c4b"} Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.097569 4934 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novaapi26a6-account-delete-ng2mb" secret="" err="secret \"galera-openstack-dockercfg-sbtdf\" not found" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.104022 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"49a440a9-1a3f-4ee9-bf28-7996fe6f222d","Type":"ContainerDied","Data":"0ae4d557135649d0f5b152d89b3a85b31e941838defa90173e4ff571d93acb6e"} Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.104125 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.129534 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder36fe-account-delete-n2bk5" podStartSLOduration=6.129511505 podStartE2EDuration="6.129511505s" podCreationTimestamp="2025-10-02 10:11:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:11:15.093243904 +0000 UTC m=+1346.845885436" watchObservedRunningTime="2025-10-02 10:11:15.129511505 +0000 UTC m=+1346.882153027" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.132043 4934 generic.go:334] "Generic (PLEG): container finished" podID="cd41dc17-6659-4b43-9758-f67d831993ba" containerID="86f78dcebecc392bfd4c80af3143e2df0ad38fe341e51b1a9c1ed70cc6f77bb2" exitCode=0 Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.132170 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-bfc5bc4b9-crmsr" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.132691 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cd41dc17-6659-4b43-9758-f67d831993ba","Type":"ContainerDied","Data":"86f78dcebecc392bfd4c80af3143e2df0ad38fe341e51b1a9c1ed70cc6f77bb2"} Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.133011 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-b97fc8746-vlx9z" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.133039 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.137070 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novaapi26a6-account-delete-ng2mb" podStartSLOduration=5.137052254 podStartE2EDuration="5.137052254s" podCreationTimestamp="2025-10-02 10:11:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:11:15.112965229 +0000 UTC m=+1346.865606751" watchObservedRunningTime="2025-10-02 10:11:15.137052254 +0000 UTC m=+1346.889693786" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.157598 4934 scope.go:117] "RemoveContainer" containerID="877a883b01af6709bd26ee32e0a326b213d51f06b683c07ca0c300ae16ad52db" Oct 02 10:11:15 crc kubenswrapper[4934]: E1002 10:11:15.158539 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"877a883b01af6709bd26ee32e0a326b213d51f06b683c07ca0c300ae16ad52db\": container with ID starting with 877a883b01af6709bd26ee32e0a326b213d51f06b683c07ca0c300ae16ad52db not found: ID does not exist" containerID="877a883b01af6709bd26ee32e0a326b213d51f06b683c07ca0c300ae16ad52db" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.158646 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"877a883b01af6709bd26ee32e0a326b213d51f06b683c07ca0c300ae16ad52db"} err="failed to get container status \"877a883b01af6709bd26ee32e0a326b213d51f06b683c07ca0c300ae16ad52db\": rpc error: code = NotFound desc = could not find container \"877a883b01af6709bd26ee32e0a326b213d51f06b683c07ca0c300ae16ad52db\": container with ID starting with 877a883b01af6709bd26ee32e0a326b213d51f06b683c07ca0c300ae16ad52db not found: ID does not exist" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.158696 4934 scope.go:117] "RemoveContainer" containerID="295f3add8c46cf850cd886f2d08991a6be7f2acfd564e68d63ce13cc92ba58df" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.189668 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cgkdl" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.200445 4934 scope.go:117] "RemoveContainer" containerID="f6b1e12bbd4d9f68b287da7ba1aa9bfc9c5284fb40a0e164b3ec1825d7817335" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.232512 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.248325 4934 scope.go:117] "RemoveContainer" containerID="38289b907213ccb98a8b74c9827dd591276a04d2139306b7c6ab58968debd430" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.250122 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.302457 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.307635 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.312876 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.317483 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.340086 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cgkdl"] Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.362080 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-bfc5bc4b9-crmsr"] Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.369120 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-bfc5bc4b9-crmsr"] Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.374217 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.377300 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-b97fc8746-vlx9z"] Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.383533 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-b97fc8746-vlx9z"] Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.390206 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.398455 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.399904 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-internal-tls-certs\") pod \"cd41dc17-6659-4b43-9758-f67d831993ba\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.399934 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cd41dc17-6659-4b43-9758-f67d831993ba-etc-machine-id\") pod \"cd41dc17-6659-4b43-9758-f67d831993ba\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.399989 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-combined-ca-bundle\") pod \"cd41dc17-6659-4b43-9758-f67d831993ba\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.400020 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7sf2w\" (UniqueName: \"kubernetes.io/projected/cd41dc17-6659-4b43-9758-f67d831993ba-kube-api-access-7sf2w\") pod \"cd41dc17-6659-4b43-9758-f67d831993ba\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.400053 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-config-data\") pod \"cd41dc17-6659-4b43-9758-f67d831993ba\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.400074 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-scripts\") pod \"cd41dc17-6659-4b43-9758-f67d831993ba\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.400124 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-config-data-custom\") pod \"cd41dc17-6659-4b43-9758-f67d831993ba\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.400193 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd41dc17-6659-4b43-9758-f67d831993ba-logs\") pod \"cd41dc17-6659-4b43-9758-f67d831993ba\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.400220 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-public-tls-certs\") pod \"cd41dc17-6659-4b43-9758-f67d831993ba\" (UID: \"cd41dc17-6659-4b43-9758-f67d831993ba\") " Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.401138 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cd41dc17-6659-4b43-9758-f67d831993ba-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "cd41dc17-6659-4b43-9758-f67d831993ba" (UID: "cd41dc17-6659-4b43-9758-f67d831993ba"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.402232 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd41dc17-6659-4b43-9758-f67d831993ba-logs" (OuterVolumeSpecName: "logs") pod "cd41dc17-6659-4b43-9758-f67d831993ba" (UID: "cd41dc17-6659-4b43-9758-f67d831993ba"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.432816 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "cd41dc17-6659-4b43-9758-f67d831993ba" (UID: "cd41dc17-6659-4b43-9758-f67d831993ba"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.432920 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-scripts" (OuterVolumeSpecName: "scripts") pod "cd41dc17-6659-4b43-9758-f67d831993ba" (UID: "cd41dc17-6659-4b43-9758-f67d831993ba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.435003 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd41dc17-6659-4b43-9758-f67d831993ba-kube-api-access-7sf2w" (OuterVolumeSpecName: "kube-api-access-7sf2w") pod "cd41dc17-6659-4b43-9758-f67d831993ba" (UID: "cd41dc17-6659-4b43-9758-f67d831993ba"). InnerVolumeSpecName "kube-api-access-7sf2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.492391 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-4z9rv"] Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.499477 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-4z9rv"] Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.502433 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd41dc17-6659-4b43-9758-f67d831993ba-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.502462 4934 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cd41dc17-6659-4b43-9758-f67d831993ba-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.502472 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7sf2w\" (UniqueName: \"kubernetes.io/projected/cd41dc17-6659-4b43-9758-f67d831993ba-kube-api-access-7sf2w\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.502482 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.503872 4934 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.510275 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi26a6-account-delete-ng2mb"] Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.514782 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-26a6-account-create-54z6g"] Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.518335 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-26a6-account-create-54z6g"] Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.535149 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-8648b56f4-5f8xk" podUID="350928d4-5a0b-4331-b1c2-220517d313a5" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.152:9311/healthcheck\": read tcp 10.217.0.2:50458->10.217.0.152:9311: read: connection reset by peer" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.535170 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-8648b56f4-5f8xk" podUID="350928d4-5a0b-4331-b1c2-220517d313a5" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.152:9311/healthcheck\": read tcp 10.217.0.2:50460->10.217.0.152:9311: read: connection reset by peer" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.535330 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "cd41dc17-6659-4b43-9758-f67d831993ba" (UID: "cd41dc17-6659-4b43-9758-f67d831993ba"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.553045 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd41dc17-6659-4b43-9758-f67d831993ba" (UID: "cd41dc17-6659-4b43-9758-f67d831993ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.572810 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-config-data" (OuterVolumeSpecName: "config-data") pod "cd41dc17-6659-4b43-9758-f67d831993ba" (UID: "cd41dc17-6659-4b43-9758-f67d831993ba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.574020 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "cd41dc17-6659-4b43-9758-f67d831993ba" (UID: "cd41dc17-6659-4b43-9758-f67d831993ba"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.605848 4934 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.605888 4934 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.605902 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.605914 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd41dc17-6659-4b43-9758-f67d831993ba-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.784255 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.798677 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.900047 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.911110 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-public-tls-certs\") pod \"ecc76715-8b05-4529-bd7b-289d7f32eff5\" (UID: \"ecc76715-8b05-4529-bd7b-289d7f32eff5\") " Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.911200 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84d2000a-d99f-4320-b178-874940ab7e9d-combined-ca-bundle\") pod \"84d2000a-d99f-4320-b178-874940ab7e9d\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.911231 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"84d2000a-d99f-4320-b178-874940ab7e9d\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.911263 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-internal-tls-certs\") pod \"ecc76715-8b05-4529-bd7b-289d7f32eff5\" (UID: \"ecc76715-8b05-4529-bd7b-289d7f32eff5\") " Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.911335 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84d2000a-d99f-4320-b178-874940ab7e9d-logs\") pod \"84d2000a-d99f-4320-b178-874940ab7e9d\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.911374 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqlxc\" (UniqueName: \"kubernetes.io/projected/84d2000a-d99f-4320-b178-874940ab7e9d-kube-api-access-zqlxc\") pod \"84d2000a-d99f-4320-b178-874940ab7e9d\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.911676 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-scripts\") pod \"ecc76715-8b05-4529-bd7b-289d7f32eff5\" (UID: \"ecc76715-8b05-4529-bd7b-289d7f32eff5\") " Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.911715 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9165cca6-da95-45ae-87c3-b2829756db3b-config-data\") pod \"9165cca6-da95-45ae-87c3-b2829756db3b\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.911756 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84d2000a-d99f-4320-b178-874940ab7e9d-scripts\") pod \"84d2000a-d99f-4320-b178-874940ab7e9d\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.911787 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecc76715-8b05-4529-bd7b-289d7f32eff5-logs\") pod \"ecc76715-8b05-4529-bd7b-289d7f32eff5\" (UID: \"ecc76715-8b05-4529-bd7b-289d7f32eff5\") " Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.911835 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-combined-ca-bundle\") pod \"ecc76715-8b05-4529-bd7b-289d7f32eff5\" (UID: \"ecc76715-8b05-4529-bd7b-289d7f32eff5\") " Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.911868 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84d2000a-d99f-4320-b178-874940ab7e9d-config-data\") pod \"84d2000a-d99f-4320-b178-874940ab7e9d\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.911921 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-config-data\") pod \"ecc76715-8b05-4529-bd7b-289d7f32eff5\" (UID: \"ecc76715-8b05-4529-bd7b-289d7f32eff5\") " Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.911960 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vk7k\" (UniqueName: \"kubernetes.io/projected/ecc76715-8b05-4529-bd7b-289d7f32eff5-kube-api-access-2vk7k\") pod \"ecc76715-8b05-4529-bd7b-289d7f32eff5\" (UID: \"ecc76715-8b05-4529-bd7b-289d7f32eff5\") " Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.911993 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9165cca6-da95-45ae-87c3-b2829756db3b-public-tls-certs\") pod \"9165cca6-da95-45ae-87c3-b2829756db3b\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.912024 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qz6fh\" (UniqueName: \"kubernetes.io/projected/9165cca6-da95-45ae-87c3-b2829756db3b-kube-api-access-qz6fh\") pod \"9165cca6-da95-45ae-87c3-b2829756db3b\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.912052 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/84d2000a-d99f-4320-b178-874940ab7e9d-httpd-run\") pod \"84d2000a-d99f-4320-b178-874940ab7e9d\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.912091 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/84d2000a-d99f-4320-b178-874940ab7e9d-internal-tls-certs\") pod \"84d2000a-d99f-4320-b178-874940ab7e9d\" (UID: \"84d2000a-d99f-4320-b178-874940ab7e9d\") " Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.912121 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9165cca6-da95-45ae-87c3-b2829756db3b-logs\") pod \"9165cca6-da95-45ae-87c3-b2829756db3b\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.913211 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9165cca6-da95-45ae-87c3-b2829756db3b-logs" (OuterVolumeSpecName: "logs") pod "9165cca6-da95-45ae-87c3-b2829756db3b" (UID: "9165cca6-da95-45ae-87c3-b2829756db3b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.916341 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "84d2000a-d99f-4320-b178-874940ab7e9d" (UID: "84d2000a-d99f-4320-b178-874940ab7e9d"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.916661 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84d2000a-d99f-4320-b178-874940ab7e9d-logs" (OuterVolumeSpecName: "logs") pod "84d2000a-d99f-4320-b178-874940ab7e9d" (UID: "84d2000a-d99f-4320-b178-874940ab7e9d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.916943 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecc76715-8b05-4529-bd7b-289d7f32eff5-logs" (OuterVolumeSpecName: "logs") pod "ecc76715-8b05-4529-bd7b-289d7f32eff5" (UID: "ecc76715-8b05-4529-bd7b-289d7f32eff5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.923484 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-scripts" (OuterVolumeSpecName: "scripts") pod "ecc76715-8b05-4529-bd7b-289d7f32eff5" (UID: "ecc76715-8b05-4529-bd7b-289d7f32eff5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.927975 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84d2000a-d99f-4320-b178-874940ab7e9d-kube-api-access-zqlxc" (OuterVolumeSpecName: "kube-api-access-zqlxc") pod "84d2000a-d99f-4320-b178-874940ab7e9d" (UID: "84d2000a-d99f-4320-b178-874940ab7e9d"). InnerVolumeSpecName "kube-api-access-zqlxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.928062 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9165cca6-da95-45ae-87c3-b2829756db3b-kube-api-access-qz6fh" (OuterVolumeSpecName: "kube-api-access-qz6fh") pod "9165cca6-da95-45ae-87c3-b2829756db3b" (UID: "9165cca6-da95-45ae-87c3-b2829756db3b"). InnerVolumeSpecName "kube-api-access-qz6fh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.939814 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84d2000a-d99f-4320-b178-874940ab7e9d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "84d2000a-d99f-4320-b178-874940ab7e9d" (UID: "84d2000a-d99f-4320-b178-874940ab7e9d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.954747 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecc76715-8b05-4529-bd7b-289d7f32eff5-kube-api-access-2vk7k" (OuterVolumeSpecName: "kube-api-access-2vk7k") pod "ecc76715-8b05-4529-bd7b-289d7f32eff5" (UID: "ecc76715-8b05-4529-bd7b-289d7f32eff5"). InnerVolumeSpecName "kube-api-access-2vk7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.954953 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84d2000a-d99f-4320-b178-874940ab7e9d-scripts" (OuterVolumeSpecName: "scripts") pod "84d2000a-d99f-4320-b178-874940ab7e9d" (UID: "84d2000a-d99f-4320-b178-874940ab7e9d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:15 crc kubenswrapper[4934]: I1002 10:11:15.959936 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84d2000a-d99f-4320-b178-874940ab7e9d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "84d2000a-d99f-4320-b178-874940ab7e9d" (UID: "84d2000a-d99f-4320-b178-874940ab7e9d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.013439 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"9165cca6-da95-45ae-87c3-b2829756db3b\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.013495 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9165cca6-da95-45ae-87c3-b2829756db3b-combined-ca-bundle\") pod \"9165cca6-da95-45ae-87c3-b2829756db3b\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.013560 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9165cca6-da95-45ae-87c3-b2829756db3b-httpd-run\") pod \"9165cca6-da95-45ae-87c3-b2829756db3b\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.013618 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9165cca6-da95-45ae-87c3-b2829756db3b-scripts\") pod \"9165cca6-da95-45ae-87c3-b2829756db3b\" (UID: \"9165cca6-da95-45ae-87c3-b2829756db3b\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.014147 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vk7k\" (UniqueName: \"kubernetes.io/projected/ecc76715-8b05-4529-bd7b-289d7f32eff5-kube-api-access-2vk7k\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.014169 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qz6fh\" (UniqueName: \"kubernetes.io/projected/9165cca6-da95-45ae-87c3-b2829756db3b-kube-api-access-qz6fh\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.014181 4934 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/84d2000a-d99f-4320-b178-874940ab7e9d-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.014192 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9165cca6-da95-45ae-87c3-b2829756db3b-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.014203 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84d2000a-d99f-4320-b178-874940ab7e9d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.014226 4934 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.014236 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84d2000a-d99f-4320-b178-874940ab7e9d-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.014249 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqlxc\" (UniqueName: \"kubernetes.io/projected/84d2000a-d99f-4320-b178-874940ab7e9d-kube-api-access-zqlxc\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.014259 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.014270 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84d2000a-d99f-4320-b178-874940ab7e9d-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.014280 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecc76715-8b05-4529-bd7b-289d7f32eff5-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.037385 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9165cca6-da95-45ae-87c3-b2829756db3b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9165cca6-da95-45ae-87c3-b2829756db3b" (UID: "9165cca6-da95-45ae-87c3-b2829756db3b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.044784 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9165cca6-da95-45ae-87c3-b2829756db3b-scripts" (OuterVolumeSpecName: "scripts") pod "9165cca6-da95-45ae-87c3-b2829756db3b" (UID: "9165cca6-da95-45ae-87c3-b2829756db3b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.044949 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "9165cca6-da95-45ae-87c3-b2829756db3b" (UID: "9165cca6-da95-45ae-87c3-b2829756db3b"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.048880 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9165cca6-da95-45ae-87c3-b2829756db3b-config-data" (OuterVolumeSpecName: "config-data") pod "9165cca6-da95-45ae-87c3-b2829756db3b" (UID: "9165cca6-da95-45ae-87c3-b2829756db3b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.051339 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84d2000a-d99f-4320-b178-874940ab7e9d-config-data" (OuterVolumeSpecName: "config-data") pod "84d2000a-d99f-4320-b178-874940ab7e9d" (UID: "84d2000a-d99f-4320-b178-874940ab7e9d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.054565 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9165cca6-da95-45ae-87c3-b2829756db3b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9165cca6-da95-45ae-87c3-b2829756db3b" (UID: "9165cca6-da95-45ae-87c3-b2829756db3b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.085335 4934 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.110080 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84d2000a-d99f-4320-b178-874940ab7e9d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "84d2000a-d99f-4320-b178-874940ab7e9d" (UID: "84d2000a-d99f-4320-b178-874940ab7e9d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.116413 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9165cca6-da95-45ae-87c3-b2829756db3b-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.116448 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84d2000a-d99f-4320-b178-874940ab7e9d-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.116460 4934 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/84d2000a-d99f-4320-b178-874940ab7e9d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.116484 4934 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.116499 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9165cca6-da95-45ae-87c3-b2829756db3b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.116511 4934 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.116522 4934 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9165cca6-da95-45ae-87c3-b2829756db3b-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.116534 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9165cca6-da95-45ae-87c3-b2829756db3b-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.121389 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.130364 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.135798 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9165cca6-da95-45ae-87c3-b2829756db3b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "9165cca6-da95-45ae-87c3-b2829756db3b" (UID: "9165cca6-da95-45ae-87c3-b2829756db3b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.139683 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.148875 4934 generic.go:334] "Generic (PLEG): container finished" podID="04aff378-509c-49be-bf4c-2d59111fc910" containerID="f28c0b65a711c22b95f0ba26fc6761c9a173fc0e9b32cd123f69c153b87e25bf" exitCode=0 Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.149016 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"04aff378-509c-49be-bf4c-2d59111fc910","Type":"ContainerDied","Data":"f28c0b65a711c22b95f0ba26fc6761c9a173fc0e9b32cd123f69c153b87e25bf"} Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.149847 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"04aff378-509c-49be-bf4c-2d59111fc910","Type":"ContainerDied","Data":"d810a75aec31cfdfe545e4469e243df734555ffbf92e9708fa55b94d75895215"} Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.149964 4934 scope.go:117] "RemoveContainer" containerID="f28c0b65a711c22b95f0ba26fc6761c9a173fc0e9b32cd123f69c153b87e25bf" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.148980 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.165045 4934 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.167861 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ecc76715-8b05-4529-bd7b-289d7f32eff5" (UID: "ecc76715-8b05-4529-bd7b-289d7f32eff5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.171102 4934 generic.go:334] "Generic (PLEG): container finished" podID="9165cca6-da95-45ae-87c3-b2829756db3b" containerID="e38759eb2a2651b29d3a6703a0da0365f7f7843aad2b4f894f68ed24320e6d21" exitCode=0 Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.171222 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9165cca6-da95-45ae-87c3-b2829756db3b","Type":"ContainerDied","Data":"e38759eb2a2651b29d3a6703a0da0365f7f7843aad2b4f894f68ed24320e6d21"} Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.171301 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9165cca6-da95-45ae-87c3-b2829756db3b","Type":"ContainerDied","Data":"76d825cb9e54693e07042d9c95aebb56e6bc394a03bc211dedb6af4cb0844a87"} Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.171415 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.179168 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ecc76715-8b05-4529-bd7b-289d7f32eff5" (UID: "ecc76715-8b05-4529-bd7b-289d7f32eff5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.185090 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-config-data" (OuterVolumeSpecName: "config-data") pod "ecc76715-8b05-4529-bd7b-289d7f32eff5" (UID: "ecc76715-8b05-4529-bd7b-289d7f32eff5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.187942 4934 generic.go:334] "Generic (PLEG): container finished" podID="350928d4-5a0b-4331-b1c2-220517d313a5" containerID="41b55b90784f0d27b65abc25b560172adc6a97c8f0800218d04b4ab6f18fc610" exitCode=0 Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.188007 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8648b56f4-5f8xk" event={"ID":"350928d4-5a0b-4331-b1c2-220517d313a5","Type":"ContainerDied","Data":"41b55b90784f0d27b65abc25b560172adc6a97c8f0800218d04b4ab6f18fc610"} Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.188030 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8648b56f4-5f8xk" event={"ID":"350928d4-5a0b-4331-b1c2-220517d313a5","Type":"ContainerDied","Data":"af20426c4cf3268848d21e7e02b8f5c6c16fdc7187bdb42848932be47df42728"} Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.188072 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8648b56f4-5f8xk" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.203424 4934 generic.go:334] "Generic (PLEG): container finished" podID="d4709e8f-24b0-48e6-ab94-39ec23f9ff64" containerID="ad980858f1e631f4751cce1464f341435468c25a526741faf7b3a1d60a45898a" exitCode=0 Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.203546 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d4709e8f-24b0-48e6-ab94-39ec23f9ff64","Type":"ContainerDied","Data":"ad980858f1e631f4751cce1464f341435468c25a526741faf7b3a1d60a45898a"} Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.206746 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.206772 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"84d2000a-d99f-4320-b178-874940ab7e9d","Type":"ContainerDied","Data":"5a03fdf3501c835a8877d0465aa7848d08e00f1009db66a43ba18f5aa66b3a00"} Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.214411 4934 generic.go:334] "Generic (PLEG): container finished" podID="ecc76715-8b05-4529-bd7b-289d7f32eff5" containerID="8710e544f71f9ca7366446c35d7eb3080fa07ae1e0f59d3a870062d5e14b9eb6" exitCode=0 Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.214563 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-649ccd9666-dh5rs" event={"ID":"ecc76715-8b05-4529-bd7b-289d7f32eff5","Type":"ContainerDied","Data":"8710e544f71f9ca7366446c35d7eb3080fa07ae1e0f59d3a870062d5e14b9eb6"} Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.214658 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-649ccd9666-dh5rs" event={"ID":"ecc76715-8b05-4529-bd7b-289d7f32eff5","Type":"ContainerDied","Data":"e4d07638d617ac5b5a4bf2034509973c9924de81e2c70de5faa9fb5370268e3c"} Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.214764 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-649ccd9666-dh5rs" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.217802 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-kolla-config\") pod \"f43ed4d4-3637-4ef6-a914-3a751cba8b4e\" (UID: \"f43ed4d4-3637-4ef6-a914-3a751cba8b4e\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.217868 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04aff378-509c-49be-bf4c-2d59111fc910-combined-ca-bundle\") pod \"04aff378-509c-49be-bf4c-2d59111fc910\" (UID: \"04aff378-509c-49be-bf4c-2d59111fc910\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.217920 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4spw\" (UniqueName: \"kubernetes.io/projected/04aff378-509c-49be-bf4c-2d59111fc910-kube-api-access-z4spw\") pod \"04aff378-509c-49be-bf4c-2d59111fc910\" (UID: \"04aff378-509c-49be-bf4c-2d59111fc910\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.217965 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-memcached-tls-certs\") pod \"f43ed4d4-3637-4ef6-a914-3a751cba8b4e\" (UID: \"f43ed4d4-3637-4ef6-a914-3a751cba8b4e\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.218103 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04aff378-509c-49be-bf4c-2d59111fc910-config-data\") pod \"04aff378-509c-49be-bf4c-2d59111fc910\" (UID: \"04aff378-509c-49be-bf4c-2d59111fc910\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.218160 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/04aff378-509c-49be-bf4c-2d59111fc910-nova-metadata-tls-certs\") pod \"04aff378-509c-49be-bf4c-2d59111fc910\" (UID: \"04aff378-509c-49be-bf4c-2d59111fc910\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.218224 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04aff378-509c-49be-bf4c-2d59111fc910-logs\") pod \"04aff378-509c-49be-bf4c-2d59111fc910\" (UID: \"04aff378-509c-49be-bf4c-2d59111fc910\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.218267 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-config-data\") pod \"f43ed4d4-3637-4ef6-a914-3a751cba8b4e\" (UID: \"f43ed4d4-3637-4ef6-a914-3a751cba8b4e\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.218294 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-combined-ca-bundle\") pod \"f43ed4d4-3637-4ef6-a914-3a751cba8b4e\" (UID: \"f43ed4d4-3637-4ef6-a914-3a751cba8b4e\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.218355 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jvtp\" (UniqueName: \"kubernetes.io/projected/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-kube-api-access-6jvtp\") pod \"f43ed4d4-3637-4ef6-a914-3a751cba8b4e\" (UID: \"f43ed4d4-3637-4ef6-a914-3a751cba8b4e\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.219547 4934 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.219571 4934 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.219651 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.219664 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.219675 4934 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9165cca6-da95-45ae-87c3-b2829756db3b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.219929 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-config-data" (OuterVolumeSpecName: "config-data") pod "f43ed4d4-3637-4ef6-a914-3a751cba8b4e" (UID: "f43ed4d4-3637-4ef6-a914-3a751cba8b4e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.220315 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04aff378-509c-49be-bf4c-2d59111fc910-logs" (OuterVolumeSpecName: "logs") pod "04aff378-509c-49be-bf4c-2d59111fc910" (UID: "04aff378-509c-49be-bf4c-2d59111fc910"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.221470 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "f43ed4d4-3637-4ef6-a914-3a751cba8b4e" (UID: "f43ed4d4-3637-4ef6-a914-3a751cba8b4e"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.230260 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-kube-api-access-6jvtp" (OuterVolumeSpecName: "kube-api-access-6jvtp") pod "f43ed4d4-3637-4ef6-a914-3a751cba8b4e" (UID: "f43ed4d4-3637-4ef6-a914-3a751cba8b4e"). InnerVolumeSpecName "kube-api-access-6jvtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.233035 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04aff378-509c-49be-bf4c-2d59111fc910-kube-api-access-z4spw" (OuterVolumeSpecName: "kube-api-access-z4spw") pod "04aff378-509c-49be-bf4c-2d59111fc910" (UID: "04aff378-509c-49be-bf4c-2d59111fc910"). InnerVolumeSpecName "kube-api-access-z4spw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.234386 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.235392 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cd41dc17-6659-4b43-9758-f67d831993ba","Type":"ContainerDied","Data":"e1c2d28a7763a945c2fdf6581a667b801498dd55a55f90a0c565eaf3b399e221"} Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.245687 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ecc76715-8b05-4529-bd7b-289d7f32eff5" (UID: "ecc76715-8b05-4529-bd7b-289d7f32eff5"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.248197 4934 generic.go:334] "Generic (PLEG): container finished" podID="f43ed4d4-3637-4ef6-a914-3a751cba8b4e" containerID="9d4708b7023ab963058c9d146e0e8bf04766c8e6b064f4e76425c8490490d0bd" exitCode=0 Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.249008 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.249197 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"f43ed4d4-3637-4ef6-a914-3a751cba8b4e","Type":"ContainerDied","Data":"9d4708b7023ab963058c9d146e0e8bf04766c8e6b064f4e76425c8490490d0bd"} Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.249294 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"f43ed4d4-3637-4ef6-a914-3a751cba8b4e","Type":"ContainerDied","Data":"0297f7f74c5f85cb4628fdc1ed774e069f7c70d754ad3ed734b372387fd7c1e2"} Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.249547 4934 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novaapi26a6-account-delete-ng2mb" secret="" err="secret \"galera-openstack-dockercfg-sbtdf\" not found" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.251407 4934 scope.go:117] "RemoveContainer" containerID="0c87adf381aab6824e14a94acbf03d404203f8ec3721590d4efd316ce48e8dd8" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.276620 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f43ed4d4-3637-4ef6-a914-3a751cba8b4e" (UID: "f43ed4d4-3637-4ef6-a914-3a751cba8b4e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.277898 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04aff378-509c-49be-bf4c-2d59111fc910-config-data" (OuterVolumeSpecName: "config-data") pod "04aff378-509c-49be-bf4c-2d59111fc910" (UID: "04aff378-509c-49be-bf4c-2d59111fc910"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.291900 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04aff378-509c-49be-bf4c-2d59111fc910-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "04aff378-509c-49be-bf4c-2d59111fc910" (UID: "04aff378-509c-49be-bf4c-2d59111fc910"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.293862 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.301022 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "f43ed4d4-3637-4ef6-a914-3a751cba8b4e" (UID: "f43ed4d4-3637-4ef6-a914-3a751cba8b4e"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.303499 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.305703 4934 scope.go:117] "RemoveContainer" containerID="f28c0b65a711c22b95f0ba26fc6761c9a173fc0e9b32cd123f69c153b87e25bf" Oct 02 10:11:16 crc kubenswrapper[4934]: E1002 10:11:16.306485 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f28c0b65a711c22b95f0ba26fc6761c9a173fc0e9b32cd123f69c153b87e25bf\": container with ID starting with f28c0b65a711c22b95f0ba26fc6761c9a173fc0e9b32cd123f69c153b87e25bf not found: ID does not exist" containerID="f28c0b65a711c22b95f0ba26fc6761c9a173fc0e9b32cd123f69c153b87e25bf" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.306521 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f28c0b65a711c22b95f0ba26fc6761c9a173fc0e9b32cd123f69c153b87e25bf"} err="failed to get container status \"f28c0b65a711c22b95f0ba26fc6761c9a173fc0e9b32cd123f69c153b87e25bf\": rpc error: code = NotFound desc = could not find container \"f28c0b65a711c22b95f0ba26fc6761c9a173fc0e9b32cd123f69c153b87e25bf\": container with ID starting with f28c0b65a711c22b95f0ba26fc6761c9a173fc0e9b32cd123f69c153b87e25bf not found: ID does not exist" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.306545 4934 scope.go:117] "RemoveContainer" containerID="0c87adf381aab6824e14a94acbf03d404203f8ec3721590d4efd316ce48e8dd8" Oct 02 10:11:16 crc kubenswrapper[4934]: E1002 10:11:16.306980 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c87adf381aab6824e14a94acbf03d404203f8ec3721590d4efd316ce48e8dd8\": container with ID starting with 0c87adf381aab6824e14a94acbf03d404203f8ec3721590d4efd316ce48e8dd8 not found: ID does not exist" containerID="0c87adf381aab6824e14a94acbf03d404203f8ec3721590d4efd316ce48e8dd8" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.307033 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c87adf381aab6824e14a94acbf03d404203f8ec3721590d4efd316ce48e8dd8"} err="failed to get container status \"0c87adf381aab6824e14a94acbf03d404203f8ec3721590d4efd316ce48e8dd8\": rpc error: code = NotFound desc = could not find container \"0c87adf381aab6824e14a94acbf03d404203f8ec3721590d4efd316ce48e8dd8\": container with ID starting with 0c87adf381aab6824e14a94acbf03d404203f8ec3721590d4efd316ce48e8dd8 not found: ID does not exist" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.307049 4934 scope.go:117] "RemoveContainer" containerID="e38759eb2a2651b29d3a6703a0da0365f7f7843aad2b4f894f68ed24320e6d21" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.312670 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.313043 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04aff378-509c-49be-bf4c-2d59111fc910-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "04aff378-509c-49be-bf4c-2d59111fc910" (UID: "04aff378-509c-49be-bf4c-2d59111fc910"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.314762 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.317675 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.320991 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmn6q\" (UniqueName: \"kubernetes.io/projected/350928d4-5a0b-4331-b1c2-220517d313a5-kube-api-access-mmn6q\") pod \"350928d4-5a0b-4331-b1c2-220517d313a5\" (UID: \"350928d4-5a0b-4331-b1c2-220517d313a5\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.321144 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-combined-ca-bundle\") pod \"350928d4-5a0b-4331-b1c2-220517d313a5\" (UID: \"350928d4-5a0b-4331-b1c2-220517d313a5\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.321220 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-config-data\") pod \"350928d4-5a0b-4331-b1c2-220517d313a5\" (UID: \"350928d4-5a0b-4331-b1c2-220517d313a5\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.321318 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/350928d4-5a0b-4331-b1c2-220517d313a5-logs\") pod \"350928d4-5a0b-4331-b1c2-220517d313a5\" (UID: \"350928d4-5a0b-4331-b1c2-220517d313a5\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.321418 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-public-tls-certs\") pod \"350928d4-5a0b-4331-b1c2-220517d313a5\" (UID: \"350928d4-5a0b-4331-b1c2-220517d313a5\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.321505 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-internal-tls-certs\") pod \"350928d4-5a0b-4331-b1c2-220517d313a5\" (UID: \"350928d4-5a0b-4331-b1c2-220517d313a5\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.321637 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-config-data-custom\") pod \"350928d4-5a0b-4331-b1c2-220517d313a5\" (UID: \"350928d4-5a0b-4331-b1c2-220517d313a5\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.322019 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.322100 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.322161 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jvtp\" (UniqueName: \"kubernetes.io/projected/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-kube-api-access-6jvtp\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.322211 4934 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.322265 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04aff378-509c-49be-bf4c-2d59111fc910-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.322314 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4spw\" (UniqueName: \"kubernetes.io/projected/04aff378-509c-49be-bf4c-2d59111fc910-kube-api-access-z4spw\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.322371 4934 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f43ed4d4-3637-4ef6-a914-3a751cba8b4e-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.322428 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04aff378-509c-49be-bf4c-2d59111fc910-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.322481 4934 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/04aff378-509c-49be-bf4c-2d59111fc910-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.322532 4934 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ecc76715-8b05-4529-bd7b-289d7f32eff5-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.322602 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04aff378-509c-49be-bf4c-2d59111fc910-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.324228 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/350928d4-5a0b-4331-b1c2-220517d313a5-logs" (OuterVolumeSpecName: "logs") pod "350928d4-5a0b-4331-b1c2-220517d313a5" (UID: "350928d4-5a0b-4331-b1c2-220517d313a5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.324354 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.328447 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "350928d4-5a0b-4331-b1c2-220517d313a5" (UID: "350928d4-5a0b-4331-b1c2-220517d313a5"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.336782 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/350928d4-5a0b-4331-b1c2-220517d313a5-kube-api-access-mmn6q" (OuterVolumeSpecName: "kube-api-access-mmn6q") pod "350928d4-5a0b-4331-b1c2-220517d313a5" (UID: "350928d4-5a0b-4331-b1c2-220517d313a5"). InnerVolumeSpecName "kube-api-access-mmn6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.338528 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.339832 4934 scope.go:117] "RemoveContainer" containerID="9dde5aaa3e65635d53137792cb4e654aa61c584c4bc197377b96acd4ff67091e" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.355399 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "350928d4-5a0b-4331-b1c2-220517d313a5" (UID: "350928d4-5a0b-4331-b1c2-220517d313a5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.376609 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "350928d4-5a0b-4331-b1c2-220517d313a5" (UID: "350928d4-5a0b-4331-b1c2-220517d313a5"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.382609 4934 scope.go:117] "RemoveContainer" containerID="e38759eb2a2651b29d3a6703a0da0365f7f7843aad2b4f894f68ed24320e6d21" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.382615 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-config-data" (OuterVolumeSpecName: "config-data") pod "350928d4-5a0b-4331-b1c2-220517d313a5" (UID: "350928d4-5a0b-4331-b1c2-220517d313a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: E1002 10:11:16.383294 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e38759eb2a2651b29d3a6703a0da0365f7f7843aad2b4f894f68ed24320e6d21\": container with ID starting with e38759eb2a2651b29d3a6703a0da0365f7f7843aad2b4f894f68ed24320e6d21 not found: ID does not exist" containerID="e38759eb2a2651b29d3a6703a0da0365f7f7843aad2b4f894f68ed24320e6d21" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.383335 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e38759eb2a2651b29d3a6703a0da0365f7f7843aad2b4f894f68ed24320e6d21"} err="failed to get container status \"e38759eb2a2651b29d3a6703a0da0365f7f7843aad2b4f894f68ed24320e6d21\": rpc error: code = NotFound desc = could not find container \"e38759eb2a2651b29d3a6703a0da0365f7f7843aad2b4f894f68ed24320e6d21\": container with ID starting with e38759eb2a2651b29d3a6703a0da0365f7f7843aad2b4f894f68ed24320e6d21 not found: ID does not exist" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.383361 4934 scope.go:117] "RemoveContainer" containerID="9dde5aaa3e65635d53137792cb4e654aa61c584c4bc197377b96acd4ff67091e" Oct 02 10:11:16 crc kubenswrapper[4934]: E1002 10:11:16.383692 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9dde5aaa3e65635d53137792cb4e654aa61c584c4bc197377b96acd4ff67091e\": container with ID starting with 9dde5aaa3e65635d53137792cb4e654aa61c584c4bc197377b96acd4ff67091e not found: ID does not exist" containerID="9dde5aaa3e65635d53137792cb4e654aa61c584c4bc197377b96acd4ff67091e" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.383716 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9dde5aaa3e65635d53137792cb4e654aa61c584c4bc197377b96acd4ff67091e"} err="failed to get container status \"9dde5aaa3e65635d53137792cb4e654aa61c584c4bc197377b96acd4ff67091e\": rpc error: code = NotFound desc = could not find container \"9dde5aaa3e65635d53137792cb4e654aa61c584c4bc197377b96acd4ff67091e\": container with ID starting with 9dde5aaa3e65635d53137792cb4e654aa61c584c4bc197377b96acd4ff67091e not found: ID does not exist" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.383733 4934 scope.go:117] "RemoveContainer" containerID="41b55b90784f0d27b65abc25b560172adc6a97c8f0800218d04b4ab6f18fc610" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.407801 4934 scope.go:117] "RemoveContainer" containerID="9245f97a19f447f9859a60a1c0d5f3f95ccff037fb5235ac35a03f447f11628f" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.421586 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-57456ffd97-rqc8b" podUID="6d79c694-2253-419e-b63b-4a38884ac57e" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.169:9696/\": dial tcp 10.217.0.169:9696: connect: connection refused" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.423397 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-public-tls-certs\") pod \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\" (UID: \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.423511 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2pdb\" (UniqueName: \"kubernetes.io/projected/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-kube-api-access-r2pdb\") pod \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\" (UID: \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.423663 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-internal-tls-certs\") pod \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\" (UID: \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.423807 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-config-data\") pod \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\" (UID: \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.423882 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-logs\") pod \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\" (UID: \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.423983 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-combined-ca-bundle\") pod \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\" (UID: \"d4709e8f-24b0-48e6-ab94-39ec23f9ff64\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.424337 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-logs" (OuterVolumeSpecName: "logs") pod "d4709e8f-24b0-48e6-ab94-39ec23f9ff64" (UID: "d4709e8f-24b0-48e6-ab94-39ec23f9ff64"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.424818 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.424918 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.424984 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.425036 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/350928d4-5a0b-4331-b1c2-220517d313a5-logs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.425086 4934 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.425140 4934 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.425190 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmn6q\" (UniqueName: \"kubernetes.io/projected/350928d4-5a0b-4331-b1c2-220517d313a5-kube-api-access-mmn6q\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.429562 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "350928d4-5a0b-4331-b1c2-220517d313a5" (UID: "350928d4-5a0b-4331-b1c2-220517d313a5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.432476 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-kube-api-access-r2pdb" (OuterVolumeSpecName: "kube-api-access-r2pdb") pod "d4709e8f-24b0-48e6-ab94-39ec23f9ff64" (UID: "d4709e8f-24b0-48e6-ab94-39ec23f9ff64"). InnerVolumeSpecName "kube-api-access-r2pdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.446851 4934 scope.go:117] "RemoveContainer" containerID="41b55b90784f0d27b65abc25b560172adc6a97c8f0800218d04b4ab6f18fc610" Oct 02 10:11:16 crc kubenswrapper[4934]: E1002 10:11:16.447440 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41b55b90784f0d27b65abc25b560172adc6a97c8f0800218d04b4ab6f18fc610\": container with ID starting with 41b55b90784f0d27b65abc25b560172adc6a97c8f0800218d04b4ab6f18fc610 not found: ID does not exist" containerID="41b55b90784f0d27b65abc25b560172adc6a97c8f0800218d04b4ab6f18fc610" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.447470 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41b55b90784f0d27b65abc25b560172adc6a97c8f0800218d04b4ab6f18fc610"} err="failed to get container status \"41b55b90784f0d27b65abc25b560172adc6a97c8f0800218d04b4ab6f18fc610\": rpc error: code = NotFound desc = could not find container \"41b55b90784f0d27b65abc25b560172adc6a97c8f0800218d04b4ab6f18fc610\": container with ID starting with 41b55b90784f0d27b65abc25b560172adc6a97c8f0800218d04b4ab6f18fc610 not found: ID does not exist" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.447494 4934 scope.go:117] "RemoveContainer" containerID="9245f97a19f447f9859a60a1c0d5f3f95ccff037fb5235ac35a03f447f11628f" Oct 02 10:11:16 crc kubenswrapper[4934]: E1002 10:11:16.447732 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9245f97a19f447f9859a60a1c0d5f3f95ccff037fb5235ac35a03f447f11628f\": container with ID starting with 9245f97a19f447f9859a60a1c0d5f3f95ccff037fb5235ac35a03f447f11628f not found: ID does not exist" containerID="9245f97a19f447f9859a60a1c0d5f3f95ccff037fb5235ac35a03f447f11628f" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.447754 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9245f97a19f447f9859a60a1c0d5f3f95ccff037fb5235ac35a03f447f11628f"} err="failed to get container status \"9245f97a19f447f9859a60a1c0d5f3f95ccff037fb5235ac35a03f447f11628f\": rpc error: code = NotFound desc = could not find container \"9245f97a19f447f9859a60a1c0d5f3f95ccff037fb5235ac35a03f447f11628f\": container with ID starting with 9245f97a19f447f9859a60a1c0d5f3f95ccff037fb5235ac35a03f447f11628f not found: ID does not exist" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.447770 4934 scope.go:117] "RemoveContainer" containerID="391ab06a0bf5924bac18ac518442b3a5539dd7de7e5001512639e7be11a54744" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.460166 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4709e8f-24b0-48e6-ab94-39ec23f9ff64" (UID: "d4709e8f-24b0-48e6-ab94-39ec23f9ff64"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.469965 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-config-data" (OuterVolumeSpecName: "config-data") pod "d4709e8f-24b0-48e6-ab94-39ec23f9ff64" (UID: "d4709e8f-24b0-48e6-ab94-39ec23f9ff64"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.479565 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.482635 4934 scope.go:117] "RemoveContainer" containerID="9799cb2e3daec429b427c098af4fc69103ad295acfa71492709948a69984306d" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.488079 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.489891 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d4709e8f-24b0-48e6-ab94-39ec23f9ff64" (UID: "d4709e8f-24b0-48e6-ab94-39ec23f9ff64"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.492406 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d4709e8f-24b0-48e6-ab94-39ec23f9ff64" (UID: "d4709e8f-24b0-48e6-ab94-39ec23f9ff64"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: E1002 10:11:16.497557 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892 is running failed: container process not found" containerID="9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 10:11:16 crc kubenswrapper[4934]: E1002 10:11:16.498197 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892 is running failed: container process not found" containerID="9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 10:11:16 crc kubenswrapper[4934]: E1002 10:11:16.498543 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892 is running failed: container process not found" containerID="9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 10:11:16 crc kubenswrapper[4934]: E1002 10:11:16.498635 4934 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-65wtc" podUID="1146cda2-78fc-40c5-b0d2-f8c19368fa2c" containerName="ovsdb-server" Oct 02 10:11:16 crc kubenswrapper[4934]: E1002 10:11:16.499486 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c63d866adb61766a772229b32b348b743c5300884ef0ae4924c680c6cb85c6e0" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 10:11:16 crc kubenswrapper[4934]: E1002 10:11:16.501410 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c63d866adb61766a772229b32b348b743c5300884ef0ae4924c680c6cb85c6e0" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 10:11:16 crc kubenswrapper[4934]: E1002 10:11:16.502712 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c63d866adb61766a772229b32b348b743c5300884ef0ae4924c680c6cb85c6e0" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 10:11:16 crc kubenswrapper[4934]: E1002 10:11:16.502772 4934 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-65wtc" podUID="1146cda2-78fc-40c5-b0d2-f8c19368fa2c" containerName="ovs-vswitchd" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.523879 4934 scope.go:117] "RemoveContainer" containerID="8710e544f71f9ca7366446c35d7eb3080fa07ae1e0f59d3a870062d5e14b9eb6" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.527102 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.542948 4934 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.543016 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2pdb\" (UniqueName: \"kubernetes.io/projected/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-kube-api-access-r2pdb\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.543035 4934 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/350928d4-5a0b-4331-b1c2-220517d313a5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.543048 4934 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.543078 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4709e8f-24b0-48e6-ab94-39ec23f9ff64-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.573028 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-8648b56f4-5f8xk"] Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.575188 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.586495 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-8648b56f4-5f8xk"] Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.593278 4934 scope.go:117] "RemoveContainer" containerID="4d19c2050c253cd17d4e337c977a53e89370af2b71d1d6898f8b9f9d88a2f2b2" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.612549 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.620782 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.631463 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-649ccd9666-dh5rs"] Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.632551 4934 scope.go:117] "RemoveContainer" containerID="8710e544f71f9ca7366446c35d7eb3080fa07ae1e0f59d3a870062d5e14b9eb6" Oct 02 10:11:16 crc kubenswrapper[4934]: E1002 10:11:16.633084 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8710e544f71f9ca7366446c35d7eb3080fa07ae1e0f59d3a870062d5e14b9eb6\": container with ID starting with 8710e544f71f9ca7366446c35d7eb3080fa07ae1e0f59d3a870062d5e14b9eb6 not found: ID does not exist" containerID="8710e544f71f9ca7366446c35d7eb3080fa07ae1e0f59d3a870062d5e14b9eb6" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.633121 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8710e544f71f9ca7366446c35d7eb3080fa07ae1e0f59d3a870062d5e14b9eb6"} err="failed to get container status \"8710e544f71f9ca7366446c35d7eb3080fa07ae1e0f59d3a870062d5e14b9eb6\": rpc error: code = NotFound desc = could not find container \"8710e544f71f9ca7366446c35d7eb3080fa07ae1e0f59d3a870062d5e14b9eb6\": container with ID starting with 8710e544f71f9ca7366446c35d7eb3080fa07ae1e0f59d3a870062d5e14b9eb6 not found: ID does not exist" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.633150 4934 scope.go:117] "RemoveContainer" containerID="4d19c2050c253cd17d4e337c977a53e89370af2b71d1d6898f8b9f9d88a2f2b2" Oct 02 10:11:16 crc kubenswrapper[4934]: E1002 10:11:16.633805 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d19c2050c253cd17d4e337c977a53e89370af2b71d1d6898f8b9f9d88a2f2b2\": container with ID starting with 4d19c2050c253cd17d4e337c977a53e89370af2b71d1d6898f8b9f9d88a2f2b2 not found: ID does not exist" containerID="4d19c2050c253cd17d4e337c977a53e89370af2b71d1d6898f8b9f9d88a2f2b2" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.633837 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d19c2050c253cd17d4e337c977a53e89370af2b71d1d6898f8b9f9d88a2f2b2"} err="failed to get container status \"4d19c2050c253cd17d4e337c977a53e89370af2b71d1d6898f8b9f9d88a2f2b2\": rpc error: code = NotFound desc = could not find container \"4d19c2050c253cd17d4e337c977a53e89370af2b71d1d6898f8b9f9d88a2f2b2\": container with ID starting with 4d19c2050c253cd17d4e337c977a53e89370af2b71d1d6898f8b9f9d88a2f2b2 not found: ID does not exist" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.633864 4934 scope.go:117] "RemoveContainer" containerID="86f78dcebecc392bfd4c80af3143e2df0ad38fe341e51b1a9c1ed70cc6f77bb2" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.637504 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-649ccd9666-dh5rs"] Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.661373 4934 scope.go:117] "RemoveContainer" containerID="ddc003d07d24a576f93d547d9f6f180b883faaf22fef13259bf076d508867ebe" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.682618 4934 scope.go:117] "RemoveContainer" containerID="9d4708b7023ab963058c9d146e0e8bf04766c8e6b064f4e76425c8490490d0bd" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.732650 4934 scope.go:117] "RemoveContainer" containerID="9d4708b7023ab963058c9d146e0e8bf04766c8e6b064f4e76425c8490490d0bd" Oct 02 10:11:16 crc kubenswrapper[4934]: E1002 10:11:16.733028 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d4708b7023ab963058c9d146e0e8bf04766c8e6b064f4e76425c8490490d0bd\": container with ID starting with 9d4708b7023ab963058c9d146e0e8bf04766c8e6b064f4e76425c8490490d0bd not found: ID does not exist" containerID="9d4708b7023ab963058c9d146e0e8bf04766c8e6b064f4e76425c8490490d0bd" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.733053 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d4708b7023ab963058c9d146e0e8bf04766c8e6b064f4e76425c8490490d0bd"} err="failed to get container status \"9d4708b7023ab963058c9d146e0e8bf04766c8e6b064f4e76425c8490490d0bd\": rpc error: code = NotFound desc = could not find container \"9d4708b7023ab963058c9d146e0e8bf04766c8e6b064f4e76425c8490490d0bd\": container with ID starting with 9d4708b7023ab963058c9d146e0e8bf04766c8e6b064f4e76425c8490490d0bd not found: ID does not exist" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.745912 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4becb20a-8bf6-458d-856e-c8d6cd43f797-run-httpd\") pod \"4becb20a-8bf6-458d-856e-c8d6cd43f797\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.746016 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-sg-core-conf-yaml\") pod \"4becb20a-8bf6-458d-856e-c8d6cd43f797\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.746089 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-config-data\") pod \"4becb20a-8bf6-458d-856e-c8d6cd43f797\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.746172 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6z6tn\" (UniqueName: \"kubernetes.io/projected/4becb20a-8bf6-458d-856e-c8d6cd43f797-kube-api-access-6z6tn\") pod \"4becb20a-8bf6-458d-856e-c8d6cd43f797\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.746239 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-ceilometer-tls-certs\") pod \"4becb20a-8bf6-458d-856e-c8d6cd43f797\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.746321 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4becb20a-8bf6-458d-856e-c8d6cd43f797-log-httpd\") pod \"4becb20a-8bf6-458d-856e-c8d6cd43f797\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.746365 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-scripts\") pod \"4becb20a-8bf6-458d-856e-c8d6cd43f797\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.746391 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-combined-ca-bundle\") pod \"4becb20a-8bf6-458d-856e-c8d6cd43f797\" (UID: \"4becb20a-8bf6-458d-856e-c8d6cd43f797\") " Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.747804 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4becb20a-8bf6-458d-856e-c8d6cd43f797-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4becb20a-8bf6-458d-856e-c8d6cd43f797" (UID: "4becb20a-8bf6-458d-856e-c8d6cd43f797"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.751657 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4becb20a-8bf6-458d-856e-c8d6cd43f797-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4becb20a-8bf6-458d-856e-c8d6cd43f797" (UID: "4becb20a-8bf6-458d-856e-c8d6cd43f797"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.751776 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-scripts" (OuterVolumeSpecName: "scripts") pod "4becb20a-8bf6-458d-856e-c8d6cd43f797" (UID: "4becb20a-8bf6-458d-856e-c8d6cd43f797"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.756761 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4becb20a-8bf6-458d-856e-c8d6cd43f797-kube-api-access-6z6tn" (OuterVolumeSpecName: "kube-api-access-6z6tn") pod "4becb20a-8bf6-458d-856e-c8d6cd43f797" (UID: "4becb20a-8bf6-458d-856e-c8d6cd43f797"). InnerVolumeSpecName "kube-api-access-6z6tn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.785280 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4becb20a-8bf6-458d-856e-c8d6cd43f797" (UID: "4becb20a-8bf6-458d-856e-c8d6cd43f797"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.805780 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "4becb20a-8bf6-458d-856e-c8d6cd43f797" (UID: "4becb20a-8bf6-458d-856e-c8d6cd43f797"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.825736 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4becb20a-8bf6-458d-856e-c8d6cd43f797" (UID: "4becb20a-8bf6-458d-856e-c8d6cd43f797"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.849224 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6z6tn\" (UniqueName: \"kubernetes.io/projected/4becb20a-8bf6-458d-856e-c8d6cd43f797-kube-api-access-6z6tn\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.849259 4934 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.849270 4934 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4becb20a-8bf6-458d-856e-c8d6cd43f797-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.849282 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.849301 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.849311 4934 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4becb20a-8bf6-458d-856e-c8d6cd43f797-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.849321 4934 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.851815 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-config-data" (OuterVolumeSpecName: "config-data") pod "4becb20a-8bf6-458d-856e-c8d6cd43f797" (UID: "4becb20a-8bf6-458d-856e-c8d6cd43f797"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.929774 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04aff378-509c-49be-bf4c-2d59111fc910" path="/var/lib/kubelet/pods/04aff378-509c-49be-bf4c-2d59111fc910/volumes" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.931193 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="125b40d7-516c-432f-ae51-703f2b18068d" path="/var/lib/kubelet/pods/125b40d7-516c-432f-ae51-703f2b18068d/volumes" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.932363 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="350928d4-5a0b-4331-b1c2-220517d313a5" path="/var/lib/kubelet/pods/350928d4-5a0b-4331-b1c2-220517d313a5/volumes" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.933936 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49a440a9-1a3f-4ee9-bf28-7996fe6f222d" path="/var/lib/kubelet/pods/49a440a9-1a3f-4ee9-bf28-7996fe6f222d/volumes" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.935104 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53b4d698-fa73-48ad-99c6-4f0aa06ea4ab" path="/var/lib/kubelet/pods/53b4d698-fa73-48ad-99c6-4f0aa06ea4ab/volumes" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.936163 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84d2000a-d99f-4320-b178-874940ab7e9d" path="/var/lib/kubelet/pods/84d2000a-d99f-4320-b178-874940ab7e9d/volumes" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.938513 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9165cca6-da95-45ae-87c3-b2829756db3b" path="/var/lib/kubelet/pods/9165cca6-da95-45ae-87c3-b2829756db3b/volumes" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.940558 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="958bc92f-fd6d-4ba0-a21f-8f1c58bac789" path="/var/lib/kubelet/pods/958bc92f-fd6d-4ba0-a21f-8f1c58bac789/volumes" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.941195 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f" path="/var/lib/kubelet/pods/9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f/volumes" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.941831 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a30daf4c-06f2-48f5-b57f-f6704fae13a9" path="/var/lib/kubelet/pods/a30daf4c-06f2-48f5-b57f-f6704fae13a9/volumes" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.943750 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd41dc17-6659-4b43-9758-f67d831993ba" path="/var/lib/kubelet/pods/cd41dc17-6659-4b43-9758-f67d831993ba/volumes" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.944564 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d15b0450-e100-4ee7-ad62-88601eedc4f2" path="/var/lib/kubelet/pods/d15b0450-e100-4ee7-ad62-88601eedc4f2/volumes" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.945785 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecc76715-8b05-4529-bd7b-289d7f32eff5" path="/var/lib/kubelet/pods/ecc76715-8b05-4529-bd7b-289d7f32eff5/volumes" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.948770 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f43ed4d4-3637-4ef6-a914-3a751cba8b4e" path="/var/lib/kubelet/pods/f43ed4d4-3637-4ef6-a914-3a751cba8b4e/volumes" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.949443 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f57c470b-0a2a-4341-ae16-6d5ac848ade0" path="/var/lib/kubelet/pods/f57c470b-0a2a-4341-ae16-6d5ac848ade0/volumes" Oct 02 10:11:16 crc kubenswrapper[4934]: I1002 10:11:16.950761 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4becb20a-8bf6-458d-856e-c8d6cd43f797-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.124184 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 02 10:11:17 crc kubenswrapper[4934]: E1002 10:11:17.202466 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6ab70322bf5ef2e1d2a4fd23bb06ac7d14bc671fa582c8397ceb83e14554def5 is running failed: container process not found" containerID="6ab70322bf5ef2e1d2a4fd23bb06ac7d14bc671fa582c8397ceb83e14554def5" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 02 10:11:17 crc kubenswrapper[4934]: E1002 10:11:17.204621 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6ab70322bf5ef2e1d2a4fd23bb06ac7d14bc671fa582c8397ceb83e14554def5 is running failed: container process not found" containerID="6ab70322bf5ef2e1d2a4fd23bb06ac7d14bc671fa582c8397ceb83e14554def5" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 02 10:11:17 crc kubenswrapper[4934]: E1002 10:11:17.205076 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6ab70322bf5ef2e1d2a4fd23bb06ac7d14bc671fa582c8397ceb83e14554def5 is running failed: container process not found" containerID="6ab70322bf5ef2e1d2a4fd23bb06ac7d14bc671fa582c8397ceb83e14554def5" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 02 10:11:17 crc kubenswrapper[4934]: E1002 10:11:17.205130 4934 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6ab70322bf5ef2e1d2a4fd23bb06ac7d14bc671fa582c8397ceb83e14554def5 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="05c8faa0-6986-415d-abc3-4d35291375f2" containerName="ovn-northd" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.257284 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.257367 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-operator-scripts\") pod \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.257400 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-secrets\") pod \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.257510 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-config-data-generated\") pod \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.257542 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76gt4\" (UniqueName: \"kubernetes.io/projected/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-kube-api-access-76gt4\") pod \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.257558 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-galera-tls-certs\") pod \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.257600 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-config-data-default\") pod \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.257625 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-combined-ca-bundle\") pod \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.257670 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-kolla-config\") pod \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\" (UID: \"8c9f23b6-92f8-4d0a-8595-84c142adc3dd\") " Oct 02 10:11:17 crc kubenswrapper[4934]: E1002 10:11:17.258002 4934 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.258028 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8c9f23b6-92f8-4d0a-8595-84c142adc3dd" (UID: "8c9f23b6-92f8-4d0a-8595-84c142adc3dd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:17 crc kubenswrapper[4934]: E1002 10:11:17.258054 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d2b6e083-a523-48e2-8157-088036174e12-config-data podName:d2b6e083-a523-48e2-8157-088036174e12 nodeName:}" failed. No retries permitted until 2025-10-02 10:11:25.258039386 +0000 UTC m=+1357.010680908 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/d2b6e083-a523-48e2-8157-088036174e12-config-data") pod "rabbitmq-cell1-server-0" (UID: "d2b6e083-a523-48e2-8157-088036174e12") : configmap "rabbitmq-cell1-config-data" not found Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.259071 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "8c9f23b6-92f8-4d0a-8595-84c142adc3dd" (UID: "8c9f23b6-92f8-4d0a-8595-84c142adc3dd"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.260221 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "8c9f23b6-92f8-4d0a-8595-84c142adc3dd" (UID: "8c9f23b6-92f8-4d0a-8595-84c142adc3dd"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.260604 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "8c9f23b6-92f8-4d0a-8595-84c142adc3dd" (UID: "8c9f23b6-92f8-4d0a-8595-84c142adc3dd"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.264634 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-secrets" (OuterVolumeSpecName: "secrets") pod "8c9f23b6-92f8-4d0a-8595-84c142adc3dd" (UID: "8c9f23b6-92f8-4d0a-8595-84c142adc3dd"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.266560 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-kube-api-access-76gt4" (OuterVolumeSpecName: "kube-api-access-76gt4") pod "8c9f23b6-92f8-4d0a-8595-84c142adc3dd" (UID: "8c9f23b6-92f8-4d0a-8595-84c142adc3dd"). InnerVolumeSpecName "kube-api-access-76gt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.270404 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "mysql-db") pod "8c9f23b6-92f8-4d0a-8595-84c142adc3dd" (UID: "8c9f23b6-92f8-4d0a-8595-84c142adc3dd"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.281119 4934 generic.go:334] "Generic (PLEG): container finished" podID="8c9f23b6-92f8-4d0a-8595-84c142adc3dd" containerID="6ed60ff6b58b2bf5ef3faeb5193e2b929c4e3c03eaae93658e8fc29218880ae3" exitCode=0 Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.281170 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8c9f23b6-92f8-4d0a-8595-84c142adc3dd","Type":"ContainerDied","Data":"6ed60ff6b58b2bf5ef3faeb5193e2b929c4e3c03eaae93658e8fc29218880ae3"} Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.281194 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8c9f23b6-92f8-4d0a-8595-84c142adc3dd","Type":"ContainerDied","Data":"612042673a7dc418a44963a83eed50002298c7097e79d4ad1bed5c687e90ef09"} Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.281210 4934 scope.go:117] "RemoveContainer" containerID="6ed60ff6b58b2bf5ef3faeb5193e2b929c4e3c03eaae93658e8fc29218880ae3" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.281302 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.282191 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c9f23b6-92f8-4d0a-8595-84c142adc3dd" (UID: "8c9f23b6-92f8-4d0a-8595-84c142adc3dd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.285839 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_05c8faa0-6986-415d-abc3-4d35291375f2/ovn-northd/0.log" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.285875 4934 generic.go:334] "Generic (PLEG): container finished" podID="05c8faa0-6986-415d-abc3-4d35291375f2" containerID="6ab70322bf5ef2e1d2a4fd23bb06ac7d14bc671fa582c8397ceb83e14554def5" exitCode=139 Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.285925 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"05c8faa0-6986-415d-abc3-4d35291375f2","Type":"ContainerDied","Data":"6ab70322bf5ef2e1d2a4fd23bb06ac7d14bc671fa582c8397ceb83e14554def5"} Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.289391 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d4709e8f-24b0-48e6-ab94-39ec23f9ff64","Type":"ContainerDied","Data":"6002a20d934ea69674bd3c416c4f994aa3e745c4fca0dd96c079749fe8cfa0ca"} Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.289491 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.326269 4934 scope.go:117] "RemoveContainer" containerID="f224fbd60ef61b586e97dd832e64710cef45baeff3c590ddb75d7e1d298abf45" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.329704 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.330932 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "8c9f23b6-92f8-4d0a-8595-84c142adc3dd" (UID: "8c9f23b6-92f8-4d0a-8595-84c142adc3dd"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.334531 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.339082 4934 generic.go:334] "Generic (PLEG): container finished" podID="4becb20a-8bf6-458d-856e-c8d6cd43f797" containerID="8e5fcc5a4c69ba03bdc17a733d109d36416615950178c66556ce847df4a0f0cf" exitCode=0 Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.339297 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/novaapi26a6-account-delete-ng2mb" podUID="a7f7a4d6-c246-4c88-85e1-573fd07ec33f" containerName="mariadb-account-delete" containerID="cri-o://4251b40d65b71246a67772b89828963553528ad5d1e49e7473e8cac466296c4b" gracePeriod=30 Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.339332 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.339560 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4becb20a-8bf6-458d-856e-c8d6cd43f797","Type":"ContainerDied","Data":"8e5fcc5a4c69ba03bdc17a733d109d36416615950178c66556ce847df4a0f0cf"} Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.339692 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4becb20a-8bf6-458d-856e-c8d6cd43f797","Type":"ContainerDied","Data":"01e10a448cab604b193449c027d3bef612ac49c44b1079e5df9e7dbe2ca1bfd1"} Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.340001 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cgkdl" podUID="029cff51-0959-4fdc-ba9e-44d61abbdf09" containerName="registry-server" containerID="cri-o://a4c2d36a689c80cbdf56b5ef00161c7b75929cbc632517514af288d8cc84f00a" gracePeriod=2 Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.363422 4934 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.363945 4934 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.363963 4934 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.365004 4934 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-secrets\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.365019 4934 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.365029 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76gt4\" (UniqueName: \"kubernetes.io/projected/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-kube-api-access-76gt4\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.365037 4934 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.365048 4934 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.365057 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c9f23b6-92f8-4d0a-8595-84c142adc3dd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.369928 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.383032 4934 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.385765 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.412390 4934 scope.go:117] "RemoveContainer" containerID="6ed60ff6b58b2bf5ef3faeb5193e2b929c4e3c03eaae93658e8fc29218880ae3" Oct 02 10:11:17 crc kubenswrapper[4934]: E1002 10:11:17.412851 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ed60ff6b58b2bf5ef3faeb5193e2b929c4e3c03eaae93658e8fc29218880ae3\": container with ID starting with 6ed60ff6b58b2bf5ef3faeb5193e2b929c4e3c03eaae93658e8fc29218880ae3 not found: ID does not exist" containerID="6ed60ff6b58b2bf5ef3faeb5193e2b929c4e3c03eaae93658e8fc29218880ae3" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.412877 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ed60ff6b58b2bf5ef3faeb5193e2b929c4e3c03eaae93658e8fc29218880ae3"} err="failed to get container status \"6ed60ff6b58b2bf5ef3faeb5193e2b929c4e3c03eaae93658e8fc29218880ae3\": rpc error: code = NotFound desc = could not find container \"6ed60ff6b58b2bf5ef3faeb5193e2b929c4e3c03eaae93658e8fc29218880ae3\": container with ID starting with 6ed60ff6b58b2bf5ef3faeb5193e2b929c4e3c03eaae93658e8fc29218880ae3 not found: ID does not exist" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.412900 4934 scope.go:117] "RemoveContainer" containerID="f224fbd60ef61b586e97dd832e64710cef45baeff3c590ddb75d7e1d298abf45" Oct 02 10:11:17 crc kubenswrapper[4934]: E1002 10:11:17.413091 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f224fbd60ef61b586e97dd832e64710cef45baeff3c590ddb75d7e1d298abf45\": container with ID starting with f224fbd60ef61b586e97dd832e64710cef45baeff3c590ddb75d7e1d298abf45 not found: ID does not exist" containerID="f224fbd60ef61b586e97dd832e64710cef45baeff3c590ddb75d7e1d298abf45" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.413106 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f224fbd60ef61b586e97dd832e64710cef45baeff3c590ddb75d7e1d298abf45"} err="failed to get container status \"f224fbd60ef61b586e97dd832e64710cef45baeff3c590ddb75d7e1d298abf45\": rpc error: code = NotFound desc = could not find container \"f224fbd60ef61b586e97dd832e64710cef45baeff3c590ddb75d7e1d298abf45\": container with ID starting with f224fbd60ef61b586e97dd832e64710cef45baeff3c590ddb75d7e1d298abf45 not found: ID does not exist" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.413119 4934 scope.go:117] "RemoveContainer" containerID="ad980858f1e631f4751cce1464f341435468c25a526741faf7b3a1d60a45898a" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.449354 4934 scope.go:117] "RemoveContainer" containerID="b19b0bf49673fda59a55fb96cbe24384993711b894ffa56c5b3e5b90372c5231" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.466874 4934 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.478940 4934 scope.go:117] "RemoveContainer" containerID="8cdeead710c71e753ff7516bcb319070ccf923a448f85b9b738b5bbc64e5abd1" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.506691 4934 scope.go:117] "RemoveContainer" containerID="1b462ec2465f66c7ac2fdf185d407ece6092286b180138bdcd246fc6359029b3" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.526070 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/keystone-7b7b4b7b85-9llg8" podUID="02feb54e-6b14-42ec-a22d-524a6005e2fd" containerName="keystone-api" probeResult="failure" output="Get \"https://10.217.0.145:5000/v3\": read tcp 10.217.0.2:50126->10.217.0.145:5000: read: connection reset by peer" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.541314 4934 scope.go:117] "RemoveContainer" containerID="8e5fcc5a4c69ba03bdc17a733d109d36416615950178c66556ce847df4a0f0cf" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.604991 4934 scope.go:117] "RemoveContainer" containerID="bf2e8afd472c84d3385087f6698782dc92f0a62388c560ef85d4823b8156cfb7" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.612076 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_05c8faa0-6986-415d-abc3-4d35291375f2/ovn-northd/0.log" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.612129 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.622683 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.626941 4934 scope.go:117] "RemoveContainer" containerID="8cdeead710c71e753ff7516bcb319070ccf923a448f85b9b738b5bbc64e5abd1" Oct 02 10:11:17 crc kubenswrapper[4934]: E1002 10:11:17.627438 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cdeead710c71e753ff7516bcb319070ccf923a448f85b9b738b5bbc64e5abd1\": container with ID starting with 8cdeead710c71e753ff7516bcb319070ccf923a448f85b9b738b5bbc64e5abd1 not found: ID does not exist" containerID="8cdeead710c71e753ff7516bcb319070ccf923a448f85b9b738b5bbc64e5abd1" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.627480 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cdeead710c71e753ff7516bcb319070ccf923a448f85b9b738b5bbc64e5abd1"} err="failed to get container status \"8cdeead710c71e753ff7516bcb319070ccf923a448f85b9b738b5bbc64e5abd1\": rpc error: code = NotFound desc = could not find container \"8cdeead710c71e753ff7516bcb319070ccf923a448f85b9b738b5bbc64e5abd1\": container with ID starting with 8cdeead710c71e753ff7516bcb319070ccf923a448f85b9b738b5bbc64e5abd1 not found: ID does not exist" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.627504 4934 scope.go:117] "RemoveContainer" containerID="1b462ec2465f66c7ac2fdf185d407ece6092286b180138bdcd246fc6359029b3" Oct 02 10:11:17 crc kubenswrapper[4934]: E1002 10:11:17.627795 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b462ec2465f66c7ac2fdf185d407ece6092286b180138bdcd246fc6359029b3\": container with ID starting with 1b462ec2465f66c7ac2fdf185d407ece6092286b180138bdcd246fc6359029b3 not found: ID does not exist" containerID="1b462ec2465f66c7ac2fdf185d407ece6092286b180138bdcd246fc6359029b3" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.627820 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b462ec2465f66c7ac2fdf185d407ece6092286b180138bdcd246fc6359029b3"} err="failed to get container status \"1b462ec2465f66c7ac2fdf185d407ece6092286b180138bdcd246fc6359029b3\": rpc error: code = NotFound desc = could not find container \"1b462ec2465f66c7ac2fdf185d407ece6092286b180138bdcd246fc6359029b3\": container with ID starting with 1b462ec2465f66c7ac2fdf185d407ece6092286b180138bdcd246fc6359029b3 not found: ID does not exist" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.627839 4934 scope.go:117] "RemoveContainer" containerID="8e5fcc5a4c69ba03bdc17a733d109d36416615950178c66556ce847df4a0f0cf" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.628191 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Oct 02 10:11:17 crc kubenswrapper[4934]: E1002 10:11:17.628228 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e5fcc5a4c69ba03bdc17a733d109d36416615950178c66556ce847df4a0f0cf\": container with ID starting with 8e5fcc5a4c69ba03bdc17a733d109d36416615950178c66556ce847df4a0f0cf not found: ID does not exist" containerID="8e5fcc5a4c69ba03bdc17a733d109d36416615950178c66556ce847df4a0f0cf" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.628271 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e5fcc5a4c69ba03bdc17a733d109d36416615950178c66556ce847df4a0f0cf"} err="failed to get container status \"8e5fcc5a4c69ba03bdc17a733d109d36416615950178c66556ce847df4a0f0cf\": rpc error: code = NotFound desc = could not find container \"8e5fcc5a4c69ba03bdc17a733d109d36416615950178c66556ce847df4a0f0cf\": container with ID starting with 8e5fcc5a4c69ba03bdc17a733d109d36416615950178c66556ce847df4a0f0cf not found: ID does not exist" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.628298 4934 scope.go:117] "RemoveContainer" containerID="bf2e8afd472c84d3385087f6698782dc92f0a62388c560ef85d4823b8156cfb7" Oct 02 10:11:17 crc kubenswrapper[4934]: E1002 10:11:17.628677 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf2e8afd472c84d3385087f6698782dc92f0a62388c560ef85d4823b8156cfb7\": container with ID starting with bf2e8afd472c84d3385087f6698782dc92f0a62388c560ef85d4823b8156cfb7 not found: ID does not exist" containerID="bf2e8afd472c84d3385087f6698782dc92f0a62388c560ef85d4823b8156cfb7" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.628704 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf2e8afd472c84d3385087f6698782dc92f0a62388c560ef85d4823b8156cfb7"} err="failed to get container status \"bf2e8afd472c84d3385087f6698782dc92f0a62388c560ef85d4823b8156cfb7\": rpc error: code = NotFound desc = could not find container \"bf2e8afd472c84d3385087f6698782dc92f0a62388c560ef85d4823b8156cfb7\": container with ID starting with bf2e8afd472c84d3385087f6698782dc92f0a62388c560ef85d4823b8156cfb7 not found: ID does not exist" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.669455 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05c8faa0-6986-415d-abc3-4d35291375f2-config\") pod \"05c8faa0-6986-415d-abc3-4d35291375f2\" (UID: \"05c8faa0-6986-415d-abc3-4d35291375f2\") " Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.669542 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/05c8faa0-6986-415d-abc3-4d35291375f2-scripts\") pod \"05c8faa0-6986-415d-abc3-4d35291375f2\" (UID: \"05c8faa0-6986-415d-abc3-4d35291375f2\") " Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.669569 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/05c8faa0-6986-415d-abc3-4d35291375f2-metrics-certs-tls-certs\") pod \"05c8faa0-6986-415d-abc3-4d35291375f2\" (UID: \"05c8faa0-6986-415d-abc3-4d35291375f2\") " Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.669609 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05c8faa0-6986-415d-abc3-4d35291375f2-combined-ca-bundle\") pod \"05c8faa0-6986-415d-abc3-4d35291375f2\" (UID: \"05c8faa0-6986-415d-abc3-4d35291375f2\") " Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.669648 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7496\" (UniqueName: \"kubernetes.io/projected/05c8faa0-6986-415d-abc3-4d35291375f2-kube-api-access-x7496\") pod \"05c8faa0-6986-415d-abc3-4d35291375f2\" (UID: \"05c8faa0-6986-415d-abc3-4d35291375f2\") " Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.669674 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/05c8faa0-6986-415d-abc3-4d35291375f2-ovn-northd-tls-certs\") pod \"05c8faa0-6986-415d-abc3-4d35291375f2\" (UID: \"05c8faa0-6986-415d-abc3-4d35291375f2\") " Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.669706 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/05c8faa0-6986-415d-abc3-4d35291375f2-ovn-rundir\") pod \"05c8faa0-6986-415d-abc3-4d35291375f2\" (UID: \"05c8faa0-6986-415d-abc3-4d35291375f2\") " Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.670318 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05c8faa0-6986-415d-abc3-4d35291375f2-scripts" (OuterVolumeSpecName: "scripts") pod "05c8faa0-6986-415d-abc3-4d35291375f2" (UID: "05c8faa0-6986-415d-abc3-4d35291375f2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.670454 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05c8faa0-6986-415d-abc3-4d35291375f2-config" (OuterVolumeSpecName: "config") pod "05c8faa0-6986-415d-abc3-4d35291375f2" (UID: "05c8faa0-6986-415d-abc3-4d35291375f2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.670474 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05c8faa0-6986-415d-abc3-4d35291375f2-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "05c8faa0-6986-415d-abc3-4d35291375f2" (UID: "05c8faa0-6986-415d-abc3-4d35291375f2"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.673612 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05c8faa0-6986-415d-abc3-4d35291375f2-kube-api-access-x7496" (OuterVolumeSpecName: "kube-api-access-x7496") pod "05c8faa0-6986-415d-abc3-4d35291375f2" (UID: "05c8faa0-6986-415d-abc3-4d35291375f2"). InnerVolumeSpecName "kube-api-access-x7496". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:17 crc kubenswrapper[4934]: E1002 10:11:17.683470 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="75db33a0a204cb84036a785aefe5e3a861f5bb1c02c563e44111dc7a54a02cb7" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 02 10:11:17 crc kubenswrapper[4934]: E1002 10:11:17.684883 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="75db33a0a204cb84036a785aefe5e3a861f5bb1c02c563e44111dc7a54a02cb7" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 02 10:11:17 crc kubenswrapper[4934]: E1002 10:11:17.686318 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="75db33a0a204cb84036a785aefe5e3a861f5bb1c02c563e44111dc7a54a02cb7" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 02 10:11:17 crc kubenswrapper[4934]: E1002 10:11:17.686357 4934 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="3dda2730-9bb9-4113-80d3-168eb5905b2f" containerName="nova-cell0-conductor-conductor" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.696527 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05c8faa0-6986-415d-abc3-4d35291375f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "05c8faa0-6986-415d-abc3-4d35291375f2" (UID: "05c8faa0-6986-415d-abc3-4d35291375f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.753212 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05c8faa0-6986-415d-abc3-4d35291375f2-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "05c8faa0-6986-415d-abc3-4d35291375f2" (UID: "05c8faa0-6986-415d-abc3-4d35291375f2"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.771479 4934 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/05c8faa0-6986-415d-abc3-4d35291375f2-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.771527 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05c8faa0-6986-415d-abc3-4d35291375f2-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.771553 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/05c8faa0-6986-415d-abc3-4d35291375f2-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.771569 4934 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/05c8faa0-6986-415d-abc3-4d35291375f2-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.771610 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05c8faa0-6986-415d-abc3-4d35291375f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.771619 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7496\" (UniqueName: \"kubernetes.io/projected/05c8faa0-6986-415d-abc3-4d35291375f2-kube-api-access-x7496\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.784504 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05c8faa0-6986-415d-abc3-4d35291375f2-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "05c8faa0-6986-415d-abc3-4d35291375f2" (UID: "05c8faa0-6986-415d-abc3-4d35291375f2"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.837387 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="d2b6e083-a523-48e2-8157-088036174e12" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.102:5671: connect: connection refused" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.873477 4934 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/05c8faa0-6986-415d-abc3-4d35291375f2-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.967922 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cgkdl" Oct 02 10:11:17 crc kubenswrapper[4934]: I1002 10:11:17.972775 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:11:17 crc kubenswrapper[4934]: E1002 10:11:17.974922 4934 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 02 10:11:17 crc kubenswrapper[4934]: E1002 10:11:17.974970 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/97560469-c459-48f4-901e-54eaded32537-config-data podName:97560469-c459-48f4-901e-54eaded32537 nodeName:}" failed. No retries permitted until 2025-10-02 10:11:25.974958817 +0000 UTC m=+1357.727600339 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/97560469-c459-48f4-901e-54eaded32537-config-data") pod "rabbitmq-server-0" (UID: "97560469-c459-48f4-901e-54eaded32537") : configmap "rabbitmq-config-data" not found Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.075974 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-combined-ca-bundle\") pod \"02feb54e-6b14-42ec-a22d-524a6005e2fd\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.076020 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/029cff51-0959-4fdc-ba9e-44d61abbdf09-catalog-content\") pod \"029cff51-0959-4fdc-ba9e-44d61abbdf09\" (UID: \"029cff51-0959-4fdc-ba9e-44d61abbdf09\") " Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.076098 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-fernet-keys\") pod \"02feb54e-6b14-42ec-a22d-524a6005e2fd\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.076124 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8cw7\" (UniqueName: \"kubernetes.io/projected/029cff51-0959-4fdc-ba9e-44d61abbdf09-kube-api-access-l8cw7\") pod \"029cff51-0959-4fdc-ba9e-44d61abbdf09\" (UID: \"029cff51-0959-4fdc-ba9e-44d61abbdf09\") " Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.076158 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-internal-tls-certs\") pod \"02feb54e-6b14-42ec-a22d-524a6005e2fd\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.076203 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-config-data\") pod \"02feb54e-6b14-42ec-a22d-524a6005e2fd\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.076224 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-scripts\") pod \"02feb54e-6b14-42ec-a22d-524a6005e2fd\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.076251 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bb5p5\" (UniqueName: \"kubernetes.io/projected/02feb54e-6b14-42ec-a22d-524a6005e2fd-kube-api-access-bb5p5\") pod \"02feb54e-6b14-42ec-a22d-524a6005e2fd\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.076270 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-public-tls-certs\") pod \"02feb54e-6b14-42ec-a22d-524a6005e2fd\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.076338 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-credential-keys\") pod \"02feb54e-6b14-42ec-a22d-524a6005e2fd\" (UID: \"02feb54e-6b14-42ec-a22d-524a6005e2fd\") " Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.076404 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/029cff51-0959-4fdc-ba9e-44d61abbdf09-utilities\") pod \"029cff51-0959-4fdc-ba9e-44d61abbdf09\" (UID: \"029cff51-0959-4fdc-ba9e-44d61abbdf09\") " Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.079771 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/029cff51-0959-4fdc-ba9e-44d61abbdf09-utilities" (OuterVolumeSpecName: "utilities") pod "029cff51-0959-4fdc-ba9e-44d61abbdf09" (UID: "029cff51-0959-4fdc-ba9e-44d61abbdf09"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.083681 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "02feb54e-6b14-42ec-a22d-524a6005e2fd" (UID: "02feb54e-6b14-42ec-a22d-524a6005e2fd"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.102863 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/029cff51-0959-4fdc-ba9e-44d61abbdf09-kube-api-access-l8cw7" (OuterVolumeSpecName: "kube-api-access-l8cw7") pod "029cff51-0959-4fdc-ba9e-44d61abbdf09" (UID: "029cff51-0959-4fdc-ba9e-44d61abbdf09"). InnerVolumeSpecName "kube-api-access-l8cw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.102921 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02feb54e-6b14-42ec-a22d-524a6005e2fd-kube-api-access-bb5p5" (OuterVolumeSpecName: "kube-api-access-bb5p5") pod "02feb54e-6b14-42ec-a22d-524a6005e2fd" (UID: "02feb54e-6b14-42ec-a22d-524a6005e2fd"). InnerVolumeSpecName "kube-api-access-bb5p5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.104494 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "02feb54e-6b14-42ec-a22d-524a6005e2fd" (UID: "02feb54e-6b14-42ec-a22d-524a6005e2fd"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.109728 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-scripts" (OuterVolumeSpecName: "scripts") pod "02feb54e-6b14-42ec-a22d-524a6005e2fd" (UID: "02feb54e-6b14-42ec-a22d-524a6005e2fd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.117105 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-config-data" (OuterVolumeSpecName: "config-data") pod "02feb54e-6b14-42ec-a22d-524a6005e2fd" (UID: "02feb54e-6b14-42ec-a22d-524a6005e2fd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.117485 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "02feb54e-6b14-42ec-a22d-524a6005e2fd" (UID: "02feb54e-6b14-42ec-a22d-524a6005e2fd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.149030 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "02feb54e-6b14-42ec-a22d-524a6005e2fd" (UID: "02feb54e-6b14-42ec-a22d-524a6005e2fd"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.149789 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "02feb54e-6b14-42ec-a22d-524a6005e2fd" (UID: "02feb54e-6b14-42ec-a22d-524a6005e2fd"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.159791 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/029cff51-0959-4fdc-ba9e-44d61abbdf09-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "029cff51-0959-4fdc-ba9e-44d61abbdf09" (UID: "029cff51-0959-4fdc-ba9e-44d61abbdf09"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.179220 4934 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.179260 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/029cff51-0959-4fdc-ba9e-44d61abbdf09-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.179274 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.179283 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/029cff51-0959-4fdc-ba9e-44d61abbdf09-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.179291 4934 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.179300 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8cw7\" (UniqueName: \"kubernetes.io/projected/029cff51-0959-4fdc-ba9e-44d61abbdf09-kube-api-access-l8cw7\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.179311 4934 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.179319 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.179326 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.179334 4934 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/02feb54e-6b14-42ec-a22d-524a6005e2fd-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.179343 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bb5p5\" (UniqueName: \"kubernetes.io/projected/02feb54e-6b14-42ec-a22d-524a6005e2fd-kube-api-access-bb5p5\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.356665 4934 generic.go:334] "Generic (PLEG): container finished" podID="02feb54e-6b14-42ec-a22d-524a6005e2fd" containerID="e7f1b0e43bf1c9bf7da08e7461ca997f5bf6711c2a32f9d4be458d1aff9e7cdc" exitCode=0 Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.356727 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7b7b4b7b85-9llg8" event={"ID":"02feb54e-6b14-42ec-a22d-524a6005e2fd","Type":"ContainerDied","Data":"e7f1b0e43bf1c9bf7da08e7461ca997f5bf6711c2a32f9d4be458d1aff9e7cdc"} Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.356752 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7b7b4b7b85-9llg8" event={"ID":"02feb54e-6b14-42ec-a22d-524a6005e2fd","Type":"ContainerDied","Data":"499eff26242dc376d8606b745f5ee91a43e209b362e362587d1c1628a3a92c05"} Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.356771 4934 scope.go:117] "RemoveContainer" containerID="e7f1b0e43bf1c9bf7da08e7461ca997f5bf6711c2a32f9d4be458d1aff9e7cdc" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.356867 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7b7b4b7b85-9llg8" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.378435 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_05c8faa0-6986-415d-abc3-4d35291375f2/ovn-northd/0.log" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.378537 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"05c8faa0-6986-415d-abc3-4d35291375f2","Type":"ContainerDied","Data":"582a5c1df61a40fae2fe36f824dd290361362b220d5f4f143379cc93d0cb24c5"} Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.378832 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.393310 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-7b7b4b7b85-9llg8"] Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.400017 4934 generic.go:334] "Generic (PLEG): container finished" podID="029cff51-0959-4fdc-ba9e-44d61abbdf09" containerID="a4c2d36a689c80cbdf56b5ef00161c7b75929cbc632517514af288d8cc84f00a" exitCode=0 Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.400308 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgkdl" event={"ID":"029cff51-0959-4fdc-ba9e-44d61abbdf09","Type":"ContainerDied","Data":"a4c2d36a689c80cbdf56b5ef00161c7b75929cbc632517514af288d8cc84f00a"} Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.400339 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cgkdl" event={"ID":"029cff51-0959-4fdc-ba9e-44d61abbdf09","Type":"ContainerDied","Data":"5c30b4bd89ca50f3b9929d6145b7904e828e61ec6c78e5bb156612d78147c723"} Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.400418 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cgkdl" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.402427 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-7b7b4b7b85-9llg8"] Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.426323 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.428763 4934 scope.go:117] "RemoveContainer" containerID="e7f1b0e43bf1c9bf7da08e7461ca997f5bf6711c2a32f9d4be458d1aff9e7cdc" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.431854 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Oct 02 10:11:18 crc kubenswrapper[4934]: E1002 10:11:18.436485 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7f1b0e43bf1c9bf7da08e7461ca997f5bf6711c2a32f9d4be458d1aff9e7cdc\": container with ID starting with e7f1b0e43bf1c9bf7da08e7461ca997f5bf6711c2a32f9d4be458d1aff9e7cdc not found: ID does not exist" containerID="e7f1b0e43bf1c9bf7da08e7461ca997f5bf6711c2a32f9d4be458d1aff9e7cdc" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.436550 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7f1b0e43bf1c9bf7da08e7461ca997f5bf6711c2a32f9d4be458d1aff9e7cdc"} err="failed to get container status \"e7f1b0e43bf1c9bf7da08e7461ca997f5bf6711c2a32f9d4be458d1aff9e7cdc\": rpc error: code = NotFound desc = could not find container \"e7f1b0e43bf1c9bf7da08e7461ca997f5bf6711c2a32f9d4be458d1aff9e7cdc\": container with ID starting with e7f1b0e43bf1c9bf7da08e7461ca997f5bf6711c2a32f9d4be458d1aff9e7cdc not found: ID does not exist" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.436610 4934 scope.go:117] "RemoveContainer" containerID="c0691673552ec68e68f4be61b4e9826705cb137abb4ac7f57faa0ea2d303510d" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.439591 4934 generic.go:334] "Generic (PLEG): container finished" podID="d2b6e083-a523-48e2-8157-088036174e12" containerID="b2d1f1b04e17b195fbd97880fd7d7c42841c576a17071e6ee86a7e804f16c194" exitCode=0 Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.439635 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d2b6e083-a523-48e2-8157-088036174e12","Type":"ContainerDied","Data":"b2d1f1b04e17b195fbd97880fd7d7c42841c576a17071e6ee86a7e804f16c194"} Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.463139 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cgkdl"] Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.468628 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cgkdl"] Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.479415 4934 scope.go:117] "RemoveContainer" containerID="6ab70322bf5ef2e1d2a4fd23bb06ac7d14bc671fa582c8397ceb83e14554def5" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.523006 4934 scope.go:117] "RemoveContainer" containerID="a4c2d36a689c80cbdf56b5ef00161c7b75929cbc632517514af288d8cc84f00a" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.547533 4934 scope.go:117] "RemoveContainer" containerID="04dbc5eb11c41f008a7566a44294731392b57cf65c3b3f5edddf43c9ef5d5cf2" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.577759 4934 scope.go:117] "RemoveContainer" containerID="d98b6d47a7c68c173a17fd3abe6da8ef4c4f1604d284e9691179197a36263ad9" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.622133 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="97560469-c459-48f4-901e-54eaded32537" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.103:5671: connect: connection refused" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.627165 4934 scope.go:117] "RemoveContainer" containerID="a4c2d36a689c80cbdf56b5ef00161c7b75929cbc632517514af288d8cc84f00a" Oct 02 10:11:18 crc kubenswrapper[4934]: E1002 10:11:18.627555 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4c2d36a689c80cbdf56b5ef00161c7b75929cbc632517514af288d8cc84f00a\": container with ID starting with a4c2d36a689c80cbdf56b5ef00161c7b75929cbc632517514af288d8cc84f00a not found: ID does not exist" containerID="a4c2d36a689c80cbdf56b5ef00161c7b75929cbc632517514af288d8cc84f00a" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.627615 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4c2d36a689c80cbdf56b5ef00161c7b75929cbc632517514af288d8cc84f00a"} err="failed to get container status \"a4c2d36a689c80cbdf56b5ef00161c7b75929cbc632517514af288d8cc84f00a\": rpc error: code = NotFound desc = could not find container \"a4c2d36a689c80cbdf56b5ef00161c7b75929cbc632517514af288d8cc84f00a\": container with ID starting with a4c2d36a689c80cbdf56b5ef00161c7b75929cbc632517514af288d8cc84f00a not found: ID does not exist" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.627641 4934 scope.go:117] "RemoveContainer" containerID="04dbc5eb11c41f008a7566a44294731392b57cf65c3b3f5edddf43c9ef5d5cf2" Oct 02 10:11:18 crc kubenswrapper[4934]: E1002 10:11:18.627910 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04dbc5eb11c41f008a7566a44294731392b57cf65c3b3f5edddf43c9ef5d5cf2\": container with ID starting with 04dbc5eb11c41f008a7566a44294731392b57cf65c3b3f5edddf43c9ef5d5cf2 not found: ID does not exist" containerID="04dbc5eb11c41f008a7566a44294731392b57cf65c3b3f5edddf43c9ef5d5cf2" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.627928 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04dbc5eb11c41f008a7566a44294731392b57cf65c3b3f5edddf43c9ef5d5cf2"} err="failed to get container status \"04dbc5eb11c41f008a7566a44294731392b57cf65c3b3f5edddf43c9ef5d5cf2\": rpc error: code = NotFound desc = could not find container \"04dbc5eb11c41f008a7566a44294731392b57cf65c3b3f5edddf43c9ef5d5cf2\": container with ID starting with 04dbc5eb11c41f008a7566a44294731392b57cf65c3b3f5edddf43c9ef5d5cf2 not found: ID does not exist" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.627942 4934 scope.go:117] "RemoveContainer" containerID="d98b6d47a7c68c173a17fd3abe6da8ef4c4f1604d284e9691179197a36263ad9" Oct 02 10:11:18 crc kubenswrapper[4934]: E1002 10:11:18.628158 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d98b6d47a7c68c173a17fd3abe6da8ef4c4f1604d284e9691179197a36263ad9\": container with ID starting with d98b6d47a7c68c173a17fd3abe6da8ef4c4f1604d284e9691179197a36263ad9 not found: ID does not exist" containerID="d98b6d47a7c68c173a17fd3abe6da8ef4c4f1604d284e9691179197a36263ad9" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.628188 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d98b6d47a7c68c173a17fd3abe6da8ef4c4f1604d284e9691179197a36263ad9"} err="failed to get container status \"d98b6d47a7c68c173a17fd3abe6da8ef4c4f1604d284e9691179197a36263ad9\": rpc error: code = NotFound desc = could not find container \"d98b6d47a7c68c173a17fd3abe6da8ef4c4f1604d284e9691179197a36263ad9\": container with ID starting with d98b6d47a7c68c173a17fd3abe6da8ef4c4f1604d284e9691179197a36263ad9 not found: ID does not exist" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.734482 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.889438 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"d2b6e083-a523-48e2-8157-088036174e12\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.889523 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d2b6e083-a523-48e2-8157-088036174e12-pod-info\") pod \"d2b6e083-a523-48e2-8157-088036174e12\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.889590 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d2b6e083-a523-48e2-8157-088036174e12-config-data\") pod \"d2b6e083-a523-48e2-8157-088036174e12\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.889664 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d2b6e083-a523-48e2-8157-088036174e12-rabbitmq-erlang-cookie\") pod \"d2b6e083-a523-48e2-8157-088036174e12\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.889937 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d2b6e083-a523-48e2-8157-088036174e12-server-conf\") pod \"d2b6e083-a523-48e2-8157-088036174e12\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.890415 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2b6e083-a523-48e2-8157-088036174e12-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "d2b6e083-a523-48e2-8157-088036174e12" (UID: "d2b6e083-a523-48e2-8157-088036174e12"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.890525 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d2b6e083-a523-48e2-8157-088036174e12-erlang-cookie-secret\") pod \"d2b6e083-a523-48e2-8157-088036174e12\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.892030 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d2b6e083-a523-48e2-8157-088036174e12-plugins-conf\") pod \"d2b6e083-a523-48e2-8157-088036174e12\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.893253 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d2b6e083-a523-48e2-8157-088036174e12-rabbitmq-confd\") pod \"d2b6e083-a523-48e2-8157-088036174e12\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.893061 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2b6e083-a523-48e2-8157-088036174e12-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "d2b6e083-a523-48e2-8157-088036174e12" (UID: "d2b6e083-a523-48e2-8157-088036174e12"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.893848 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d2b6e083-a523-48e2-8157-088036174e12-rabbitmq-tls\") pod \"d2b6e083-a523-48e2-8157-088036174e12\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.894076 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/d2b6e083-a523-48e2-8157-088036174e12-pod-info" (OuterVolumeSpecName: "pod-info") pod "d2b6e083-a523-48e2-8157-088036174e12" (UID: "d2b6e083-a523-48e2-8157-088036174e12"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.894279 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mqz7\" (UniqueName: \"kubernetes.io/projected/d2b6e083-a523-48e2-8157-088036174e12-kube-api-access-6mqz7\") pod \"d2b6e083-a523-48e2-8157-088036174e12\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.894794 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2b6e083-a523-48e2-8157-088036174e12-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "d2b6e083-a523-48e2-8157-088036174e12" (UID: "d2b6e083-a523-48e2-8157-088036174e12"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.894919 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d2b6e083-a523-48e2-8157-088036174e12-rabbitmq-plugins\") pod \"d2b6e083-a523-48e2-8157-088036174e12\" (UID: \"d2b6e083-a523-48e2-8157-088036174e12\") " Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.895401 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2b6e083-a523-48e2-8157-088036174e12-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "d2b6e083-a523-48e2-8157-088036174e12" (UID: "d2b6e083-a523-48e2-8157-088036174e12"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.895716 4934 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d2b6e083-a523-48e2-8157-088036174e12-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.895732 4934 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d2b6e083-a523-48e2-8157-088036174e12-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.895743 4934 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d2b6e083-a523-48e2-8157-088036174e12-pod-info\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.895753 4934 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d2b6e083-a523-48e2-8157-088036174e12-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.895766 4934 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d2b6e083-a523-48e2-8157-088036174e12-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.895992 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2b6e083-a523-48e2-8157-088036174e12-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "d2b6e083-a523-48e2-8157-088036174e12" (UID: "d2b6e083-a523-48e2-8157-088036174e12"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.898888 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2b6e083-a523-48e2-8157-088036174e12-kube-api-access-6mqz7" (OuterVolumeSpecName: "kube-api-access-6mqz7") pod "d2b6e083-a523-48e2-8157-088036174e12" (UID: "d2b6e083-a523-48e2-8157-088036174e12"). InnerVolumeSpecName "kube-api-access-6mqz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.900717 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "d2b6e083-a523-48e2-8157-088036174e12" (UID: "d2b6e083-a523-48e2-8157-088036174e12"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.932476 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2b6e083-a523-48e2-8157-088036174e12-config-data" (OuterVolumeSpecName: "config-data") pod "d2b6e083-a523-48e2-8157-088036174e12" (UID: "d2b6e083-a523-48e2-8157-088036174e12"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.954518 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="029cff51-0959-4fdc-ba9e-44d61abbdf09" path="/var/lib/kubelet/pods/029cff51-0959-4fdc-ba9e-44d61abbdf09/volumes" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.955739 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02feb54e-6b14-42ec-a22d-524a6005e2fd" path="/var/lib/kubelet/pods/02feb54e-6b14-42ec-a22d-524a6005e2fd/volumes" Oct 02 10:11:18 crc kubenswrapper[4934]: E1002 10:11:18.956215 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b7f8b726c37d4fd8d67700b1306d05906985514bcb56092fd5f3e2779cf2dc8a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.956912 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05c8faa0-6986-415d-abc3-4d35291375f2" path="/var/lib/kubelet/pods/05c8faa0-6986-415d-abc3-4d35291375f2/volumes" Oct 02 10:11:18 crc kubenswrapper[4934]: E1002 10:11:18.957743 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b7f8b726c37d4fd8d67700b1306d05906985514bcb56092fd5f3e2779cf2dc8a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.958933 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4becb20a-8bf6-458d-856e-c8d6cd43f797" path="/var/lib/kubelet/pods/4becb20a-8bf6-458d-856e-c8d6cd43f797/volumes" Oct 02 10:11:18 crc kubenswrapper[4934]: E1002 10:11:18.961900 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b7f8b726c37d4fd8d67700b1306d05906985514bcb56092fd5f3e2779cf2dc8a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 10:11:18 crc kubenswrapper[4934]: E1002 10:11:18.961940 4934 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="06adab43-0177-440c-aeef-9386345160c4" containerName="nova-scheduler-scheduler" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.967118 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c9f23b6-92f8-4d0a-8595-84c142adc3dd" path="/var/lib/kubelet/pods/8c9f23b6-92f8-4d0a-8595-84c142adc3dd/volumes" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.968295 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4709e8f-24b0-48e6-ab94-39ec23f9ff64" path="/var/lib/kubelet/pods/d4709e8f-24b0-48e6-ab94-39ec23f9ff64/volumes" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.997078 4934 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.997117 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d2b6e083-a523-48e2-8157-088036174e12-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.997130 4934 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d2b6e083-a523-48e2-8157-088036174e12-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:18 crc kubenswrapper[4934]: I1002 10:11:18.997145 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mqz7\" (UniqueName: \"kubernetes.io/projected/d2b6e083-a523-48e2-8157-088036174e12-kube-api-access-6mqz7\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.006262 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2b6e083-a523-48e2-8157-088036174e12-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "d2b6e083-a523-48e2-8157-088036174e12" (UID: "d2b6e083-a523-48e2-8157-088036174e12"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.010253 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2b6e083-a523-48e2-8157-088036174e12-server-conf" (OuterVolumeSpecName: "server-conf") pod "d2b6e083-a523-48e2-8157-088036174e12" (UID: "d2b6e083-a523-48e2-8157-088036174e12"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.017042 4934 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.107185 4934 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d2b6e083-a523-48e2-8157-088036174e12-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.107234 4934 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.107244 4934 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d2b6e083-a523-48e2-8157-088036174e12-server-conf\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.236928 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.309389 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/97560469-c459-48f4-901e-54eaded32537-rabbitmq-erlang-cookie\") pod \"97560469-c459-48f4-901e-54eaded32537\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.309436 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/97560469-c459-48f4-901e-54eaded32537-erlang-cookie-secret\") pod \"97560469-c459-48f4-901e-54eaded32537\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.309491 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sk6pk\" (UniqueName: \"kubernetes.io/projected/97560469-c459-48f4-901e-54eaded32537-kube-api-access-sk6pk\") pod \"97560469-c459-48f4-901e-54eaded32537\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.309563 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/97560469-c459-48f4-901e-54eaded32537-rabbitmq-plugins\") pod \"97560469-c459-48f4-901e-54eaded32537\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.309635 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/97560469-c459-48f4-901e-54eaded32537-config-data\") pod \"97560469-c459-48f4-901e-54eaded32537\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.309678 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/97560469-c459-48f4-901e-54eaded32537-server-conf\") pod \"97560469-c459-48f4-901e-54eaded32537\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.309712 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/97560469-c459-48f4-901e-54eaded32537-plugins-conf\") pod \"97560469-c459-48f4-901e-54eaded32537\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.309732 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/97560469-c459-48f4-901e-54eaded32537-rabbitmq-tls\") pod \"97560469-c459-48f4-901e-54eaded32537\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.309755 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/97560469-c459-48f4-901e-54eaded32537-pod-info\") pod \"97560469-c459-48f4-901e-54eaded32537\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.309776 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"97560469-c459-48f4-901e-54eaded32537\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.309801 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/97560469-c459-48f4-901e-54eaded32537-rabbitmq-confd\") pod \"97560469-c459-48f4-901e-54eaded32537\" (UID: \"97560469-c459-48f4-901e-54eaded32537\") " Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.309859 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97560469-c459-48f4-901e-54eaded32537-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "97560469-c459-48f4-901e-54eaded32537" (UID: "97560469-c459-48f4-901e-54eaded32537"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.310021 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97560469-c459-48f4-901e-54eaded32537-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "97560469-c459-48f4-901e-54eaded32537" (UID: "97560469-c459-48f4-901e-54eaded32537"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.310333 4934 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/97560469-c459-48f4-901e-54eaded32537-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.310353 4934 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/97560469-c459-48f4-901e-54eaded32537-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.310515 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97560469-c459-48f4-901e-54eaded32537-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "97560469-c459-48f4-901e-54eaded32537" (UID: "97560469-c459-48f4-901e-54eaded32537"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.319748 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/97560469-c459-48f4-901e-54eaded32537-pod-info" (OuterVolumeSpecName: "pod-info") pod "97560469-c459-48f4-901e-54eaded32537" (UID: "97560469-c459-48f4-901e-54eaded32537"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.319790 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "persistence") pod "97560469-c459-48f4-901e-54eaded32537" (UID: "97560469-c459-48f4-901e-54eaded32537"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.319836 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97560469-c459-48f4-901e-54eaded32537-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "97560469-c459-48f4-901e-54eaded32537" (UID: "97560469-c459-48f4-901e-54eaded32537"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.319843 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97560469-c459-48f4-901e-54eaded32537-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "97560469-c459-48f4-901e-54eaded32537" (UID: "97560469-c459-48f4-901e-54eaded32537"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.319882 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97560469-c459-48f4-901e-54eaded32537-kube-api-access-sk6pk" (OuterVolumeSpecName: "kube-api-access-sk6pk") pod "97560469-c459-48f4-901e-54eaded32537" (UID: "97560469-c459-48f4-901e-54eaded32537"). InnerVolumeSpecName "kube-api-access-sk6pk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.332601 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97560469-c459-48f4-901e-54eaded32537-config-data" (OuterVolumeSpecName: "config-data") pod "97560469-c459-48f4-901e-54eaded32537" (UID: "97560469-c459-48f4-901e-54eaded32537"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.359048 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97560469-c459-48f4-901e-54eaded32537-server-conf" (OuterVolumeSpecName: "server-conf") pod "97560469-c459-48f4-901e-54eaded32537" (UID: "97560469-c459-48f4-901e-54eaded32537"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.397387 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97560469-c459-48f4-901e-54eaded32537-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "97560469-c459-48f4-901e-54eaded32537" (UID: "97560469-c459-48f4-901e-54eaded32537"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.411967 4934 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/97560469-c459-48f4-901e-54eaded32537-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.412013 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sk6pk\" (UniqueName: \"kubernetes.io/projected/97560469-c459-48f4-901e-54eaded32537-kube-api-access-sk6pk\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.412028 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/97560469-c459-48f4-901e-54eaded32537-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.412040 4934 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/97560469-c459-48f4-901e-54eaded32537-server-conf\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.412051 4934 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/97560469-c459-48f4-901e-54eaded32537-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.412062 4934 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/97560469-c459-48f4-901e-54eaded32537-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.412073 4934 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/97560469-c459-48f4-901e-54eaded32537-pod-info\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.412108 4934 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.412119 4934 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/97560469-c459-48f4-901e-54eaded32537-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.447072 4934 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.452825 4934 generic.go:334] "Generic (PLEG): container finished" podID="97560469-c459-48f4-901e-54eaded32537" containerID="2432f321bd3453c2b78b0dac285ee73ee7b8165fefef31515a2963451230f5a8" exitCode=0 Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.452884 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"97560469-c459-48f4-901e-54eaded32537","Type":"ContainerDied","Data":"2432f321bd3453c2b78b0dac285ee73ee7b8165fefef31515a2963451230f5a8"} Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.452945 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"97560469-c459-48f4-901e-54eaded32537","Type":"ContainerDied","Data":"d45171f7a1d489fdaf8f0fc59fb7c0b5a8d0522ee54658cd489af97652b84e4b"} Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.452965 4934 scope.go:117] "RemoveContainer" containerID="2432f321bd3453c2b78b0dac285ee73ee7b8165fefef31515a2963451230f5a8" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.453226 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.459590 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d2b6e083-a523-48e2-8157-088036174e12","Type":"ContainerDied","Data":"035be76bc25d2b585700ddf5bdb5f97eee1b3039ef484bb89d5941763c64ad15"} Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.459666 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.474734 4934 scope.go:117] "RemoveContainer" containerID="91a99821c085c0ca3aac996ae178fce6d284a5d09eab41f0c4748d65f0d4756d" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.499527 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.516860 4934 scope.go:117] "RemoveContainer" containerID="2432f321bd3453c2b78b0dac285ee73ee7b8165fefef31515a2963451230f5a8" Oct 02 10:11:19 crc kubenswrapper[4934]: E1002 10:11:19.517333 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2432f321bd3453c2b78b0dac285ee73ee7b8165fefef31515a2963451230f5a8\": container with ID starting with 2432f321bd3453c2b78b0dac285ee73ee7b8165fefef31515a2963451230f5a8 not found: ID does not exist" containerID="2432f321bd3453c2b78b0dac285ee73ee7b8165fefef31515a2963451230f5a8" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.517372 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2432f321bd3453c2b78b0dac285ee73ee7b8165fefef31515a2963451230f5a8"} err="failed to get container status \"2432f321bd3453c2b78b0dac285ee73ee7b8165fefef31515a2963451230f5a8\": rpc error: code = NotFound desc = could not find container \"2432f321bd3453c2b78b0dac285ee73ee7b8165fefef31515a2963451230f5a8\": container with ID starting with 2432f321bd3453c2b78b0dac285ee73ee7b8165fefef31515a2963451230f5a8 not found: ID does not exist" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.517396 4934 scope.go:117] "RemoveContainer" containerID="91a99821c085c0ca3aac996ae178fce6d284a5d09eab41f0c4748d65f0d4756d" Oct 02 10:11:19 crc kubenswrapper[4934]: E1002 10:11:19.517722 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91a99821c085c0ca3aac996ae178fce6d284a5d09eab41f0c4748d65f0d4756d\": container with ID starting with 91a99821c085c0ca3aac996ae178fce6d284a5d09eab41f0c4748d65f0d4756d not found: ID does not exist" containerID="91a99821c085c0ca3aac996ae178fce6d284a5d09eab41f0c4748d65f0d4756d" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.517770 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91a99821c085c0ca3aac996ae178fce6d284a5d09eab41f0c4748d65f0d4756d"} err="failed to get container status \"91a99821c085c0ca3aac996ae178fce6d284a5d09eab41f0c4748d65f0d4756d\": rpc error: code = NotFound desc = could not find container \"91a99821c085c0ca3aac996ae178fce6d284a5d09eab41f0c4748d65f0d4756d\": container with ID starting with 91a99821c085c0ca3aac996ae178fce6d284a5d09eab41f0c4748d65f0d4756d not found: ID does not exist" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.517795 4934 scope.go:117] "RemoveContainer" containerID="b2d1f1b04e17b195fbd97880fd7d7c42841c576a17071e6ee86a7e804f16c194" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.518221 4934 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.523314 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.528652 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.534192 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 10:11:19 crc kubenswrapper[4934]: I1002 10:11:19.534545 4934 scope.go:117] "RemoveContainer" containerID="eaa7cefa7a244d7ebc3161825f4958eb21752c3953ea843f3cde9449733066de" Oct 02 10:11:20 crc kubenswrapper[4934]: I1002 10:11:20.481064 4934 generic.go:334] "Generic (PLEG): container finished" podID="06adab43-0177-440c-aeef-9386345160c4" containerID="b7f8b726c37d4fd8d67700b1306d05906985514bcb56092fd5f3e2779cf2dc8a" exitCode=0 Oct 02 10:11:20 crc kubenswrapper[4934]: I1002 10:11:20.481146 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"06adab43-0177-440c-aeef-9386345160c4","Type":"ContainerDied","Data":"b7f8b726c37d4fd8d67700b1306d05906985514bcb56092fd5f3e2779cf2dc8a"} Oct 02 10:11:20 crc kubenswrapper[4934]: I1002 10:11:20.483742 4934 generic.go:334] "Generic (PLEG): container finished" podID="3dda2730-9bb9-4113-80d3-168eb5905b2f" containerID="75db33a0a204cb84036a785aefe5e3a861f5bb1c02c563e44111dc7a54a02cb7" exitCode=0 Oct 02 10:11:20 crc kubenswrapper[4934]: I1002 10:11:20.483777 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"3dda2730-9bb9-4113-80d3-168eb5905b2f","Type":"ContainerDied","Data":"75db33a0a204cb84036a785aefe5e3a861f5bb1c02c563e44111dc7a54a02cb7"} Oct 02 10:11:20 crc kubenswrapper[4934]: I1002 10:11:20.650782 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 10:11:20 crc kubenswrapper[4934]: I1002 10:11:20.658434 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 10:11:20 crc kubenswrapper[4934]: I1002 10:11:20.836462 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dda2730-9bb9-4113-80d3-168eb5905b2f-config-data\") pod \"3dda2730-9bb9-4113-80d3-168eb5905b2f\" (UID: \"3dda2730-9bb9-4113-80d3-168eb5905b2f\") " Oct 02 10:11:20 crc kubenswrapper[4934]: I1002 10:11:20.836546 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfd24\" (UniqueName: \"kubernetes.io/projected/06adab43-0177-440c-aeef-9386345160c4-kube-api-access-pfd24\") pod \"06adab43-0177-440c-aeef-9386345160c4\" (UID: \"06adab43-0177-440c-aeef-9386345160c4\") " Oct 02 10:11:20 crc kubenswrapper[4934]: I1002 10:11:20.836628 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06adab43-0177-440c-aeef-9386345160c4-combined-ca-bundle\") pod \"06adab43-0177-440c-aeef-9386345160c4\" (UID: \"06adab43-0177-440c-aeef-9386345160c4\") " Oct 02 10:11:20 crc kubenswrapper[4934]: I1002 10:11:20.836703 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlnd5\" (UniqueName: \"kubernetes.io/projected/3dda2730-9bb9-4113-80d3-168eb5905b2f-kube-api-access-rlnd5\") pod \"3dda2730-9bb9-4113-80d3-168eb5905b2f\" (UID: \"3dda2730-9bb9-4113-80d3-168eb5905b2f\") " Oct 02 10:11:20 crc kubenswrapper[4934]: I1002 10:11:20.836758 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06adab43-0177-440c-aeef-9386345160c4-config-data\") pod \"06adab43-0177-440c-aeef-9386345160c4\" (UID: \"06adab43-0177-440c-aeef-9386345160c4\") " Oct 02 10:11:20 crc kubenswrapper[4934]: I1002 10:11:20.836789 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dda2730-9bb9-4113-80d3-168eb5905b2f-combined-ca-bundle\") pod \"3dda2730-9bb9-4113-80d3-168eb5905b2f\" (UID: \"3dda2730-9bb9-4113-80d3-168eb5905b2f\") " Oct 02 10:11:20 crc kubenswrapper[4934]: I1002 10:11:20.854865 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06adab43-0177-440c-aeef-9386345160c4-kube-api-access-pfd24" (OuterVolumeSpecName: "kube-api-access-pfd24") pod "06adab43-0177-440c-aeef-9386345160c4" (UID: "06adab43-0177-440c-aeef-9386345160c4"). InnerVolumeSpecName "kube-api-access-pfd24". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:20 crc kubenswrapper[4934]: I1002 10:11:20.854973 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dda2730-9bb9-4113-80d3-168eb5905b2f-kube-api-access-rlnd5" (OuterVolumeSpecName: "kube-api-access-rlnd5") pod "3dda2730-9bb9-4113-80d3-168eb5905b2f" (UID: "3dda2730-9bb9-4113-80d3-168eb5905b2f"). InnerVolumeSpecName "kube-api-access-rlnd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:20 crc kubenswrapper[4934]: I1002 10:11:20.868794 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06adab43-0177-440c-aeef-9386345160c4-config-data" (OuterVolumeSpecName: "config-data") pod "06adab43-0177-440c-aeef-9386345160c4" (UID: "06adab43-0177-440c-aeef-9386345160c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:20 crc kubenswrapper[4934]: I1002 10:11:20.868909 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06adab43-0177-440c-aeef-9386345160c4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06adab43-0177-440c-aeef-9386345160c4" (UID: "06adab43-0177-440c-aeef-9386345160c4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:20 crc kubenswrapper[4934]: I1002 10:11:20.885453 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dda2730-9bb9-4113-80d3-168eb5905b2f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3dda2730-9bb9-4113-80d3-168eb5905b2f" (UID: "3dda2730-9bb9-4113-80d3-168eb5905b2f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:20 crc kubenswrapper[4934]: I1002 10:11:20.888366 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dda2730-9bb9-4113-80d3-168eb5905b2f-config-data" (OuterVolumeSpecName: "config-data") pod "3dda2730-9bb9-4113-80d3-168eb5905b2f" (UID: "3dda2730-9bb9-4113-80d3-168eb5905b2f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:20 crc kubenswrapper[4934]: I1002 10:11:20.924231 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97560469-c459-48f4-901e-54eaded32537" path="/var/lib/kubelet/pods/97560469-c459-48f4-901e-54eaded32537/volumes" Oct 02 10:11:20 crc kubenswrapper[4934]: I1002 10:11:20.925169 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2b6e083-a523-48e2-8157-088036174e12" path="/var/lib/kubelet/pods/d2b6e083-a523-48e2-8157-088036174e12/volumes" Oct 02 10:11:20 crc kubenswrapper[4934]: I1002 10:11:20.938236 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dda2730-9bb9-4113-80d3-168eb5905b2f-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:20 crc kubenswrapper[4934]: I1002 10:11:20.938270 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfd24\" (UniqueName: \"kubernetes.io/projected/06adab43-0177-440c-aeef-9386345160c4-kube-api-access-pfd24\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:20 crc kubenswrapper[4934]: I1002 10:11:20.938280 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06adab43-0177-440c-aeef-9386345160c4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:20 crc kubenswrapper[4934]: I1002 10:11:20.938290 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlnd5\" (UniqueName: \"kubernetes.io/projected/3dda2730-9bb9-4113-80d3-168eb5905b2f-kube-api-access-rlnd5\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:20 crc kubenswrapper[4934]: I1002 10:11:20.938298 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06adab43-0177-440c-aeef-9386345160c4-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:20 crc kubenswrapper[4934]: I1002 10:11:20.938306 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dda2730-9bb9-4113-80d3-168eb5905b2f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:21 crc kubenswrapper[4934]: I1002 10:11:21.493798 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 10:11:21 crc kubenswrapper[4934]: I1002 10:11:21.493793 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"3dda2730-9bb9-4113-80d3-168eb5905b2f","Type":"ContainerDied","Data":"397bdf51f4e9d02cbc9b0c8f8eee6ce13158d4ffeeb8f8a82b3db7867e57f59b"} Oct 02 10:11:21 crc kubenswrapper[4934]: I1002 10:11:21.493982 4934 scope.go:117] "RemoveContainer" containerID="75db33a0a204cb84036a785aefe5e3a861f5bb1c02c563e44111dc7a54a02cb7" Oct 02 10:11:21 crc kubenswrapper[4934]: E1002 10:11:21.497088 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892 is running failed: container process not found" containerID="9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 10:11:21 crc kubenswrapper[4934]: I1002 10:11:21.497169 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"06adab43-0177-440c-aeef-9386345160c4","Type":"ContainerDied","Data":"617ae8bc456aa9d23f135d10b5fd03a0b75adf6e6fb088e0d1779fd1c7dc57ba"} Oct 02 10:11:21 crc kubenswrapper[4934]: I1002 10:11:21.497262 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 10:11:21 crc kubenswrapper[4934]: E1002 10:11:21.497451 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892 is running failed: container process not found" containerID="9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 10:11:21 crc kubenswrapper[4934]: E1002 10:11:21.497714 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892 is running failed: container process not found" containerID="9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 10:11:21 crc kubenswrapper[4934]: E1002 10:11:21.497742 4934 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-65wtc" podUID="1146cda2-78fc-40c5-b0d2-f8c19368fa2c" containerName="ovsdb-server" Oct 02 10:11:21 crc kubenswrapper[4934]: E1002 10:11:21.500354 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c63d866adb61766a772229b32b348b743c5300884ef0ae4924c680c6cb85c6e0" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 10:11:21 crc kubenswrapper[4934]: E1002 10:11:21.502596 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c63d866adb61766a772229b32b348b743c5300884ef0ae4924c680c6cb85c6e0" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 10:11:21 crc kubenswrapper[4934]: E1002 10:11:21.505801 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c63d866adb61766a772229b32b348b743c5300884ef0ae4924c680c6cb85c6e0" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 10:11:21 crc kubenswrapper[4934]: E1002 10:11:21.505850 4934 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-65wtc" podUID="1146cda2-78fc-40c5-b0d2-f8c19368fa2c" containerName="ovs-vswitchd" Oct 02 10:11:21 crc kubenswrapper[4934]: I1002 10:11:21.521636 4934 scope.go:117] "RemoveContainer" containerID="b7f8b726c37d4fd8d67700b1306d05906985514bcb56092fd5f3e2779cf2dc8a" Oct 02 10:11:21 crc kubenswrapper[4934]: I1002 10:11:21.528029 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 10:11:21 crc kubenswrapper[4934]: I1002 10:11:21.539147 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 10:11:21 crc kubenswrapper[4934]: I1002 10:11:21.546939 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:11:21 crc kubenswrapper[4934]: I1002 10:11:21.553518 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 10:11:22 crc kubenswrapper[4934]: I1002 10:11:22.925149 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06adab43-0177-440c-aeef-9386345160c4" path="/var/lib/kubelet/pods/06adab43-0177-440c-aeef-9386345160c4/volumes" Oct 02 10:11:22 crc kubenswrapper[4934]: I1002 10:11:22.926143 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dda2730-9bb9-4113-80d3-168eb5905b2f" path="/var/lib/kubelet/pods/3dda2730-9bb9-4113-80d3-168eb5905b2f/volumes" Oct 02 10:11:26 crc kubenswrapper[4934]: E1002 10:11:26.497558 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892 is running failed: container process not found" containerID="9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 10:11:26 crc kubenswrapper[4934]: E1002 10:11:26.498499 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892 is running failed: container process not found" containerID="9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 10:11:26 crc kubenswrapper[4934]: E1002 10:11:26.499002 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c63d866adb61766a772229b32b348b743c5300884ef0ae4924c680c6cb85c6e0" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 10:11:26 crc kubenswrapper[4934]: E1002 10:11:26.499070 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892 is running failed: container process not found" containerID="9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 10:11:26 crc kubenswrapper[4934]: E1002 10:11:26.499089 4934 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-65wtc" podUID="1146cda2-78fc-40c5-b0d2-f8c19368fa2c" containerName="ovsdb-server" Oct 02 10:11:26 crc kubenswrapper[4934]: E1002 10:11:26.500785 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c63d866adb61766a772229b32b348b743c5300884ef0ae4924c680c6cb85c6e0" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 10:11:26 crc kubenswrapper[4934]: E1002 10:11:26.502531 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c63d866adb61766a772229b32b348b743c5300884ef0ae4924c680c6cb85c6e0" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 10:11:26 crc kubenswrapper[4934]: E1002 10:11:26.502595 4934 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-65wtc" podUID="1146cda2-78fc-40c5-b0d2-f8c19368fa2c" containerName="ovs-vswitchd" Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.397908 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-57456ffd97-rqc8b" Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.531734 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-internal-tls-certs\") pod \"6d79c694-2253-419e-b63b-4a38884ac57e\" (UID: \"6d79c694-2253-419e-b63b-4a38884ac57e\") " Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.531848 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95sg5\" (UniqueName: \"kubernetes.io/projected/6d79c694-2253-419e-b63b-4a38884ac57e-kube-api-access-95sg5\") pod \"6d79c694-2253-419e-b63b-4a38884ac57e\" (UID: \"6d79c694-2253-419e-b63b-4a38884ac57e\") " Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.531873 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-combined-ca-bundle\") pod \"6d79c694-2253-419e-b63b-4a38884ac57e\" (UID: \"6d79c694-2253-419e-b63b-4a38884ac57e\") " Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.531931 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-httpd-config\") pod \"6d79c694-2253-419e-b63b-4a38884ac57e\" (UID: \"6d79c694-2253-419e-b63b-4a38884ac57e\") " Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.531970 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-config\") pod \"6d79c694-2253-419e-b63b-4a38884ac57e\" (UID: \"6d79c694-2253-419e-b63b-4a38884ac57e\") " Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.531994 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-ovndb-tls-certs\") pod \"6d79c694-2253-419e-b63b-4a38884ac57e\" (UID: \"6d79c694-2253-419e-b63b-4a38884ac57e\") " Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.532075 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-public-tls-certs\") pod \"6d79c694-2253-419e-b63b-4a38884ac57e\" (UID: \"6d79c694-2253-419e-b63b-4a38884ac57e\") " Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.549737 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "6d79c694-2253-419e-b63b-4a38884ac57e" (UID: "6d79c694-2253-419e-b63b-4a38884ac57e"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.549787 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d79c694-2253-419e-b63b-4a38884ac57e-kube-api-access-95sg5" (OuterVolumeSpecName: "kube-api-access-95sg5") pod "6d79c694-2253-419e-b63b-4a38884ac57e" (UID: "6d79c694-2253-419e-b63b-4a38884ac57e"). InnerVolumeSpecName "kube-api-access-95sg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.573438 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6d79c694-2253-419e-b63b-4a38884ac57e" (UID: "6d79c694-2253-419e-b63b-4a38884ac57e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.575100 4934 generic.go:334] "Generic (PLEG): container finished" podID="6d79c694-2253-419e-b63b-4a38884ac57e" containerID="a68b3c26edd8aa7fbcd688303c8a83effaa456e62c9b48088363c448ed90aef6" exitCode=0 Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.575153 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-57456ffd97-rqc8b" event={"ID":"6d79c694-2253-419e-b63b-4a38884ac57e","Type":"ContainerDied","Data":"a68b3c26edd8aa7fbcd688303c8a83effaa456e62c9b48088363c448ed90aef6"} Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.575186 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-57456ffd97-rqc8b" event={"ID":"6d79c694-2253-419e-b63b-4a38884ac57e","Type":"ContainerDied","Data":"868a863f516906649063d12e1987b02363fef789aab5c137eb1a8f1869fc16a2"} Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.575225 4934 scope.go:117] "RemoveContainer" containerID="959724f7a775ac28fa7ae269f3be79b301f1332c985d6f8a80a47def488f67f5" Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.575407 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-57456ffd97-rqc8b" Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.580008 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6d79c694-2253-419e-b63b-4a38884ac57e" (UID: "6d79c694-2253-419e-b63b-4a38884ac57e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.604984 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6d79c694-2253-419e-b63b-4a38884ac57e" (UID: "6d79c694-2253-419e-b63b-4a38884ac57e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.609390 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "6d79c694-2253-419e-b63b-4a38884ac57e" (UID: "6d79c694-2253-419e-b63b-4a38884ac57e"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.609922 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-config" (OuterVolumeSpecName: "config") pod "6d79c694-2253-419e-b63b-4a38884ac57e" (UID: "6d79c694-2253-419e-b63b-4a38884ac57e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.611749 4934 scope.go:117] "RemoveContainer" containerID="a68b3c26edd8aa7fbcd688303c8a83effaa456e62c9b48088363c448ed90aef6" Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.634596 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.634637 4934 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.634647 4934 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.634657 4934 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.634666 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95sg5\" (UniqueName: \"kubernetes.io/projected/6d79c694-2253-419e-b63b-4a38884ac57e-kube-api-access-95sg5\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.634677 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.634687 4934 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6d79c694-2253-419e-b63b-4a38884ac57e-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.639455 4934 scope.go:117] "RemoveContainer" containerID="959724f7a775ac28fa7ae269f3be79b301f1332c985d6f8a80a47def488f67f5" Oct 02 10:11:27 crc kubenswrapper[4934]: E1002 10:11:27.639890 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"959724f7a775ac28fa7ae269f3be79b301f1332c985d6f8a80a47def488f67f5\": container with ID starting with 959724f7a775ac28fa7ae269f3be79b301f1332c985d6f8a80a47def488f67f5 not found: ID does not exist" containerID="959724f7a775ac28fa7ae269f3be79b301f1332c985d6f8a80a47def488f67f5" Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.639919 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"959724f7a775ac28fa7ae269f3be79b301f1332c985d6f8a80a47def488f67f5"} err="failed to get container status \"959724f7a775ac28fa7ae269f3be79b301f1332c985d6f8a80a47def488f67f5\": rpc error: code = NotFound desc = could not find container \"959724f7a775ac28fa7ae269f3be79b301f1332c985d6f8a80a47def488f67f5\": container with ID starting with 959724f7a775ac28fa7ae269f3be79b301f1332c985d6f8a80a47def488f67f5 not found: ID does not exist" Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.639940 4934 scope.go:117] "RemoveContainer" containerID="a68b3c26edd8aa7fbcd688303c8a83effaa456e62c9b48088363c448ed90aef6" Oct 02 10:11:27 crc kubenswrapper[4934]: E1002 10:11:27.640287 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a68b3c26edd8aa7fbcd688303c8a83effaa456e62c9b48088363c448ed90aef6\": container with ID starting with a68b3c26edd8aa7fbcd688303c8a83effaa456e62c9b48088363c448ed90aef6 not found: ID does not exist" containerID="a68b3c26edd8aa7fbcd688303c8a83effaa456e62c9b48088363c448ed90aef6" Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.640309 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a68b3c26edd8aa7fbcd688303c8a83effaa456e62c9b48088363c448ed90aef6"} err="failed to get container status \"a68b3c26edd8aa7fbcd688303c8a83effaa456e62c9b48088363c448ed90aef6\": rpc error: code = NotFound desc = could not find container \"a68b3c26edd8aa7fbcd688303c8a83effaa456e62c9b48088363c448ed90aef6\": container with ID starting with a68b3c26edd8aa7fbcd688303c8a83effaa456e62c9b48088363c448ed90aef6 not found: ID does not exist" Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.917921 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-57456ffd97-rqc8b"] Oct 02 10:11:27 crc kubenswrapper[4934]: I1002 10:11:27.926866 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-57456ffd97-rqc8b"] Oct 02 10:11:28 crc kubenswrapper[4934]: I1002 10:11:28.927945 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d79c694-2253-419e-b63b-4a38884ac57e" path="/var/lib/kubelet/pods/6d79c694-2253-419e-b63b-4a38884ac57e/volumes" Oct 02 10:11:31 crc kubenswrapper[4934]: E1002 10:11:31.497632 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892 is running failed: container process not found" containerID="9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 10:11:31 crc kubenswrapper[4934]: E1002 10:11:31.498781 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892 is running failed: container process not found" containerID="9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 10:11:31 crc kubenswrapper[4934]: E1002 10:11:31.499479 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892 is running failed: container process not found" containerID="9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 10:11:31 crc kubenswrapper[4934]: E1002 10:11:31.499519 4934 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-65wtc" podUID="1146cda2-78fc-40c5-b0d2-f8c19368fa2c" containerName="ovsdb-server" Oct 02 10:11:31 crc kubenswrapper[4934]: E1002 10:11:31.499779 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c63d866adb61766a772229b32b348b743c5300884ef0ae4924c680c6cb85c6e0" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 10:11:31 crc kubenswrapper[4934]: E1002 10:11:31.502548 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c63d866adb61766a772229b32b348b743c5300884ef0ae4924c680c6cb85c6e0" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 10:11:31 crc kubenswrapper[4934]: E1002 10:11:31.505482 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c63d866adb61766a772229b32b348b743c5300884ef0ae4924c680c6cb85c6e0" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 10:11:31 crc kubenswrapper[4934]: E1002 10:11:31.505531 4934 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-65wtc" podUID="1146cda2-78fc-40c5-b0d2-f8c19368fa2c" containerName="ovs-vswitchd" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.453558 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-77cj7"] Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.454555 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2b6e083-a523-48e2-8157-088036174e12" containerName="setup-container" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.454591 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2b6e083-a523-48e2-8157-088036174e12" containerName="setup-container" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.454610 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dda2730-9bb9-4113-80d3-168eb5905b2f" containerName="nova-cell0-conductor-conductor" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.454618 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dda2730-9bb9-4113-80d3-168eb5905b2f" containerName="nova-cell0-conductor-conductor" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.454639 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2b6e083-a523-48e2-8157-088036174e12" containerName="rabbitmq" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.454646 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2b6e083-a523-48e2-8157-088036174e12" containerName="rabbitmq" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.454654 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d15b0450-e100-4ee7-ad62-88601eedc4f2" containerName="barbican-keystone-listener" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.454660 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="d15b0450-e100-4ee7-ad62-88601eedc4f2" containerName="barbican-keystone-listener" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.454670 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20b39d07-1559-4d11-b9ff-6ef97142a58a" containerName="dnsmasq-dns" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.454677 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="20b39d07-1559-4d11-b9ff-6ef97142a58a" containerName="dnsmasq-dns" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.454688 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d79c694-2253-419e-b63b-4a38884ac57e" containerName="neutron-api" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.454695 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d79c694-2253-419e-b63b-4a38884ac57e" containerName="neutron-api" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.454704 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d2000a-d99f-4320-b178-874940ab7e9d" containerName="glance-httpd" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.454710 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d2000a-d99f-4320-b178-874940ab7e9d" containerName="glance-httpd" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.454720 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd41dc17-6659-4b43-9758-f67d831993ba" containerName="cinder-api" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.454727 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd41dc17-6659-4b43-9758-f67d831993ba" containerName="cinder-api" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.454735 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="029cff51-0959-4fdc-ba9e-44d61abbdf09" containerName="extract-utilities" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.454741 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="029cff51-0959-4fdc-ba9e-44d61abbdf09" containerName="extract-utilities" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.454750 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a30daf4c-06f2-48f5-b57f-f6704fae13a9" containerName="proxy-httpd" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.454755 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a30daf4c-06f2-48f5-b57f-f6704fae13a9" containerName="proxy-httpd" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.454766 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4709e8f-24b0-48e6-ab94-39ec23f9ff64" containerName="nova-api-log" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.454772 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4709e8f-24b0-48e6-ab94-39ec23f9ff64" containerName="nova-api-log" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.454782 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="350928d4-5a0b-4331-b1c2-220517d313a5" containerName="barbican-api" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.454787 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="350928d4-5a0b-4331-b1c2-220517d313a5" containerName="barbican-api" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.454794 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d2000a-d99f-4320-b178-874940ab7e9d" containerName="glance-log" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.454800 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d2000a-d99f-4320-b178-874940ab7e9d" containerName="glance-log" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.454808 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="547c6c6d-0f01-4feb-acc6-2e3045407b64" containerName="barbican-worker-log" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.454814 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="547c6c6d-0f01-4feb-acc6-2e3045407b64" containerName="barbican-worker-log" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.454824 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02feb54e-6b14-42ec-a22d-524a6005e2fd" containerName="keystone-api" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.454830 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="02feb54e-6b14-42ec-a22d-524a6005e2fd" containerName="keystone-api" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.454836 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecc76715-8b05-4529-bd7b-289d7f32eff5" containerName="placement-log" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.454842 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecc76715-8b05-4529-bd7b-289d7f32eff5" containerName="placement-log" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.454852 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4709e8f-24b0-48e6-ab94-39ec23f9ff64" containerName="nova-api-api" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.454857 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4709e8f-24b0-48e6-ab94-39ec23f9ff64" containerName="nova-api-api" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.454864 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04aff378-509c-49be-bf4c-2d59111fc910" containerName="nova-metadata-log" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.454870 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="04aff378-509c-49be-bf4c-2d59111fc910" containerName="nova-metadata-log" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.454882 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97560469-c459-48f4-901e-54eaded32537" containerName="rabbitmq" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.454887 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="97560469-c459-48f4-901e-54eaded32537" containerName="rabbitmq" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.454895 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4becb20a-8bf6-458d-856e-c8d6cd43f797" containerName="ceilometer-notification-agent" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.454901 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="4becb20a-8bf6-458d-856e-c8d6cd43f797" containerName="ceilometer-notification-agent" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.454912 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05c8faa0-6986-415d-abc3-4d35291375f2" containerName="ovn-northd" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.454918 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="05c8faa0-6986-415d-abc3-4d35291375f2" containerName="ovn-northd" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.454925 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04aff378-509c-49be-bf4c-2d59111fc910" containerName="nova-metadata-metadata" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.454930 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="04aff378-509c-49be-bf4c-2d59111fc910" containerName="nova-metadata-metadata" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.454940 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e14738be-f83f-47ce-b0d9-ceb23d504b46" containerName="openstack-network-exporter" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.454945 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="e14738be-f83f-47ce-b0d9-ceb23d504b46" containerName="openstack-network-exporter" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.454954 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20b39d07-1559-4d11-b9ff-6ef97142a58a" containerName="init" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.454960 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="20b39d07-1559-4d11-b9ff-6ef97142a58a" containerName="init" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.454970 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f43ed4d4-3637-4ef6-a914-3a751cba8b4e" containerName="memcached" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.454975 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f43ed4d4-3637-4ef6-a914-3a751cba8b4e" containerName="memcached" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.454985 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6061002c-7242-4d02-b868-6838b39f30fa" containerName="openstack-network-exporter" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.454991 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="6061002c-7242-4d02-b868-6838b39f30fa" containerName="openstack-network-exporter" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.455001 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4becb20a-8bf6-458d-856e-c8d6cd43f797" containerName="ceilometer-central-agent" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455007 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="4becb20a-8bf6-458d-856e-c8d6cd43f797" containerName="ceilometer-central-agent" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.455016 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d79c694-2253-419e-b63b-4a38884ac57e" containerName="neutron-httpd" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455022 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d79c694-2253-419e-b63b-4a38884ac57e" containerName="neutron-httpd" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.455031 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="547c6c6d-0f01-4feb-acc6-2e3045407b64" containerName="barbican-worker" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455036 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="547c6c6d-0f01-4feb-acc6-2e3045407b64" containerName="barbican-worker" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.455043 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd41dc17-6659-4b43-9758-f67d831993ba" containerName="cinder-api-log" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455049 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd41dc17-6659-4b43-9758-f67d831993ba" containerName="cinder-api-log" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.455056 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="350928d4-5a0b-4331-b1c2-220517d313a5" containerName="barbican-api-log" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455061 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="350928d4-5a0b-4331-b1c2-220517d313a5" containerName="barbican-api-log" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.455073 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05c8faa0-6986-415d-abc3-4d35291375f2" containerName="openstack-network-exporter" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455079 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="05c8faa0-6986-415d-abc3-4d35291375f2" containerName="openstack-network-exporter" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.455088 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49a440a9-1a3f-4ee9-bf28-7996fe6f222d" containerName="nova-cell1-conductor-conductor" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455094 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="49a440a9-1a3f-4ee9-bf28-7996fe6f222d" containerName="nova-cell1-conductor-conductor" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.455102 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c9f23b6-92f8-4d0a-8595-84c142adc3dd" containerName="galera" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455107 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c9f23b6-92f8-4d0a-8595-84c142adc3dd" containerName="galera" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.455116 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06adab43-0177-440c-aeef-9386345160c4" containerName="nova-scheduler-scheduler" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455122 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="06adab43-0177-440c-aeef-9386345160c4" containerName="nova-scheduler-scheduler" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.455131 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c9f23b6-92f8-4d0a-8595-84c142adc3dd" containerName="mysql-bootstrap" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455137 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c9f23b6-92f8-4d0a-8595-84c142adc3dd" containerName="mysql-bootstrap" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.455145 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecc76715-8b05-4529-bd7b-289d7f32eff5" containerName="placement-api" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455150 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecc76715-8b05-4529-bd7b-289d7f32eff5" containerName="placement-api" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.455158 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="958bc92f-fd6d-4ba0-a21f-8f1c58bac789" containerName="kube-state-metrics" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455164 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="958bc92f-fd6d-4ba0-a21f-8f1c58bac789" containerName="kube-state-metrics" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.455173 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7214291c-3c8e-4173-a6fd-ef16170258e2" containerName="ovn-controller" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455179 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="7214291c-3c8e-4173-a6fd-ef16170258e2" containerName="ovn-controller" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.455187 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a30daf4c-06f2-48f5-b57f-f6704fae13a9" containerName="proxy-server" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455193 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a30daf4c-06f2-48f5-b57f-f6704fae13a9" containerName="proxy-server" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.455203 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="029cff51-0959-4fdc-ba9e-44d61abbdf09" containerName="registry-server" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455209 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="029cff51-0959-4fdc-ba9e-44d61abbdf09" containerName="registry-server" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.455217 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d15b0450-e100-4ee7-ad62-88601eedc4f2" containerName="barbican-keystone-listener-log" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455223 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="d15b0450-e100-4ee7-ad62-88601eedc4f2" containerName="barbican-keystone-listener-log" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.455231 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455237 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.455246 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4becb20a-8bf6-458d-856e-c8d6cd43f797" containerName="proxy-httpd" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455251 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="4becb20a-8bf6-458d-856e-c8d6cd43f797" containerName="proxy-httpd" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.455260 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97560469-c459-48f4-901e-54eaded32537" containerName="setup-container" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455265 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="97560469-c459-48f4-901e-54eaded32537" containerName="setup-container" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.455271 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6061002c-7242-4d02-b868-6838b39f30fa" containerName="ovsdbserver-sb" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455277 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="6061002c-7242-4d02-b868-6838b39f30fa" containerName="ovsdbserver-sb" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.455288 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9165cca6-da95-45ae-87c3-b2829756db3b" containerName="glance-log" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455294 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="9165cca6-da95-45ae-87c3-b2829756db3b" containerName="glance-log" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.455299 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9165cca6-da95-45ae-87c3-b2829756db3b" containerName="glance-httpd" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455304 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="9165cca6-da95-45ae-87c3-b2829756db3b" containerName="glance-httpd" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.455314 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="125b40d7-516c-432f-ae51-703f2b18068d" containerName="galera" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455319 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="125b40d7-516c-432f-ae51-703f2b18068d" containerName="galera" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.455329 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="125b40d7-516c-432f-ae51-703f2b18068d" containerName="mysql-bootstrap" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455335 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="125b40d7-516c-432f-ae51-703f2b18068d" containerName="mysql-bootstrap" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.455345 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d816754-2d4c-40e6-acee-218947ab9e19" containerName="openstack-network-exporter" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455350 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d816754-2d4c-40e6-acee-218947ab9e19" containerName="openstack-network-exporter" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.455359 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d816754-2d4c-40e6-acee-218947ab9e19" containerName="ovsdbserver-nb" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455364 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d816754-2d4c-40e6-acee-218947ab9e19" containerName="ovsdbserver-nb" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.455372 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4becb20a-8bf6-458d-856e-c8d6cd43f797" containerName="sg-core" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455377 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="4becb20a-8bf6-458d-856e-c8d6cd43f797" containerName="sg-core" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.455382 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="029cff51-0959-4fdc-ba9e-44d61abbdf09" containerName="extract-content" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455388 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="029cff51-0959-4fdc-ba9e-44d61abbdf09" containerName="extract-content" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455525 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4709e8f-24b0-48e6-ab94-39ec23f9ff64" containerName="nova-api-api" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455539 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d816754-2d4c-40e6-acee-218947ab9e19" containerName="ovsdbserver-nb" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455544 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="350928d4-5a0b-4331-b1c2-220517d313a5" containerName="barbican-api" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455551 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="4becb20a-8bf6-458d-856e-c8d6cd43f797" containerName="proxy-httpd" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455558 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="547c6c6d-0f01-4feb-acc6-2e3045407b64" containerName="barbican-worker" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455567 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="125b40d7-516c-432f-ae51-703f2b18068d" containerName="galera" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455595 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd41dc17-6659-4b43-9758-f67d831993ba" containerName="cinder-api-log" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455609 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="49a440a9-1a3f-4ee9-bf28-7996fe6f222d" containerName="nova-cell1-conductor-conductor" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455621 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="029cff51-0959-4fdc-ba9e-44d61abbdf09" containerName="registry-server" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455632 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="97560469-c459-48f4-901e-54eaded32537" containerName="rabbitmq" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455639 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="958bc92f-fd6d-4ba0-a21f-8f1c58bac789" containerName="kube-state-metrics" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455655 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dda2730-9bb9-4113-80d3-168eb5905b2f" containerName="nova-cell0-conductor-conductor" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455667 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="7214291c-3c8e-4173-a6fd-ef16170258e2" containerName="ovn-controller" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455679 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="6061002c-7242-4d02-b868-6838b39f30fa" containerName="openstack-network-exporter" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455689 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="4becb20a-8bf6-458d-856e-c8d6cd43f797" containerName="ceilometer-central-agent" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455698 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="a30daf4c-06f2-48f5-b57f-f6704fae13a9" containerName="proxy-server" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455706 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="05c8faa0-6986-415d-abc3-4d35291375f2" containerName="openstack-network-exporter" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455716 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d79c694-2253-419e-b63b-4a38884ac57e" containerName="neutron-httpd" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455722 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="6061002c-7242-4d02-b868-6838b39f30fa" containerName="ovsdbserver-sb" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455729 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="9165cca6-da95-45ae-87c3-b2829756db3b" containerName="glance-log" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455737 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="20b39d07-1559-4d11-b9ff-6ef97142a58a" containerName="dnsmasq-dns" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455749 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="02feb54e-6b14-42ec-a22d-524a6005e2fd" containerName="keystone-api" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455756 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="350928d4-5a0b-4331-b1c2-220517d313a5" containerName="barbican-api-log" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455765 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="04aff378-509c-49be-bf4c-2d59111fc910" containerName="nova-metadata-log" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455774 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="d15b0450-e100-4ee7-ad62-88601eedc4f2" containerName="barbican-keystone-listener-log" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455783 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d79c694-2253-419e-b63b-4a38884ac57e" containerName="neutron-api" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455793 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f43ed4d4-3637-4ef6-a914-3a751cba8b4e" containerName="memcached" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455803 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="a30daf4c-06f2-48f5-b57f-f6704fae13a9" containerName="proxy-httpd" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455809 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="547c6c6d-0f01-4feb-acc6-2e3045407b64" containerName="barbican-worker-log" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455817 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="84d2000a-d99f-4320-b178-874940ab7e9d" containerName="glance-log" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455827 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="4becb20a-8bf6-458d-856e-c8d6cd43f797" containerName="ceilometer-notification-agent" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455837 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d816754-2d4c-40e6-acee-218947ab9e19" containerName="openstack-network-exporter" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455845 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="9165cca6-da95-45ae-87c3-b2829756db3b" containerName="glance-httpd" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455854 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4709e8f-24b0-48e6-ab94-39ec23f9ff64" containerName="nova-api-log" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455860 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="d15b0450-e100-4ee7-ad62-88601eedc4f2" containerName="barbican-keystone-listener" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455868 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="06adab43-0177-440c-aeef-9386345160c4" containerName="nova-scheduler-scheduler" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455878 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecc76715-8b05-4529-bd7b-289d7f32eff5" containerName="placement-api" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455885 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="04aff378-509c-49be-bf4c-2d59111fc910" containerName="nova-metadata-metadata" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455893 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="84d2000a-d99f-4320-b178-874940ab7e9d" containerName="glance-httpd" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455902 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2b6e083-a523-48e2-8157-088036174e12" containerName="rabbitmq" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455912 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="e14738be-f83f-47ce-b0d9-ceb23d504b46" containerName="openstack-network-exporter" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455922 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="05c8faa0-6986-415d-abc3-4d35291375f2" containerName="ovn-northd" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455931 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c9f23b6-92f8-4d0a-8595-84c142adc3dd" containerName="galera" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455939 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f37c71e-afcc-44c5-aa1b-bf8fb6d8af8f" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455946 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="4becb20a-8bf6-458d-856e-c8d6cd43f797" containerName="sg-core" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455955 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd41dc17-6659-4b43-9758-f67d831993ba" containerName="cinder-api" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.455960 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecc76715-8b05-4529-bd7b-289d7f32eff5" containerName="placement-log" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.457159 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-77cj7" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.470682 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-77cj7"] Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.518728 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffkfq\" (UniqueName: \"kubernetes.io/projected/cfb4baab-8e79-44ed-8d2a-7da87c069cc2-kube-api-access-ffkfq\") pod \"redhat-operators-77cj7\" (UID: \"cfb4baab-8e79-44ed-8d2a-7da87c069cc2\") " pod="openshift-marketplace/redhat-operators-77cj7" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.518737 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892 is running failed: container process not found" containerID="9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.518863 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfb4baab-8e79-44ed-8d2a-7da87c069cc2-catalog-content\") pod \"redhat-operators-77cj7\" (UID: \"cfb4baab-8e79-44ed-8d2a-7da87c069cc2\") " pod="openshift-marketplace/redhat-operators-77cj7" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.518924 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfb4baab-8e79-44ed-8d2a-7da87c069cc2-utilities\") pod \"redhat-operators-77cj7\" (UID: \"cfb4baab-8e79-44ed-8d2a-7da87c069cc2\") " pod="openshift-marketplace/redhat-operators-77cj7" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.519968 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892 is running failed: container process not found" containerID="9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.520312 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892 is running failed: container process not found" containerID="9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.520351 4934 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-65wtc" podUID="1146cda2-78fc-40c5-b0d2-f8c19368fa2c" containerName="ovsdb-server" Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.527547 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c63d866adb61766a772229b32b348b743c5300884ef0ae4924c680c6cb85c6e0" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.529303 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c63d866adb61766a772229b32b348b743c5300884ef0ae4924c680c6cb85c6e0" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.540101 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c63d866adb61766a772229b32b348b743c5300884ef0ae4924c680c6cb85c6e0" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 10:11:36 crc kubenswrapper[4934]: E1002 10:11:36.540235 4934 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-65wtc" podUID="1146cda2-78fc-40c5-b0d2-f8c19368fa2c" containerName="ovs-vswitchd" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.620201 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffkfq\" (UniqueName: \"kubernetes.io/projected/cfb4baab-8e79-44ed-8d2a-7da87c069cc2-kube-api-access-ffkfq\") pod \"redhat-operators-77cj7\" (UID: \"cfb4baab-8e79-44ed-8d2a-7da87c069cc2\") " pod="openshift-marketplace/redhat-operators-77cj7" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.620349 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfb4baab-8e79-44ed-8d2a-7da87c069cc2-catalog-content\") pod \"redhat-operators-77cj7\" (UID: \"cfb4baab-8e79-44ed-8d2a-7da87c069cc2\") " pod="openshift-marketplace/redhat-operators-77cj7" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.620413 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfb4baab-8e79-44ed-8d2a-7da87c069cc2-utilities\") pod \"redhat-operators-77cj7\" (UID: \"cfb4baab-8e79-44ed-8d2a-7da87c069cc2\") " pod="openshift-marketplace/redhat-operators-77cj7" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.621152 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfb4baab-8e79-44ed-8d2a-7da87c069cc2-utilities\") pod \"redhat-operators-77cj7\" (UID: \"cfb4baab-8e79-44ed-8d2a-7da87c069cc2\") " pod="openshift-marketplace/redhat-operators-77cj7" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.621218 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfb4baab-8e79-44ed-8d2a-7da87c069cc2-catalog-content\") pod \"redhat-operators-77cj7\" (UID: \"cfb4baab-8e79-44ed-8d2a-7da87c069cc2\") " pod="openshift-marketplace/redhat-operators-77cj7" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.639651 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffkfq\" (UniqueName: \"kubernetes.io/projected/cfb4baab-8e79-44ed-8d2a-7da87c069cc2-kube-api-access-ffkfq\") pod \"redhat-operators-77cj7\" (UID: \"cfb4baab-8e79-44ed-8d2a-7da87c069cc2\") " pod="openshift-marketplace/redhat-operators-77cj7" Oct 02 10:11:36 crc kubenswrapper[4934]: I1002 10:11:36.828596 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-77cj7" Oct 02 10:11:37 crc kubenswrapper[4934]: I1002 10:11:37.265460 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-77cj7"] Oct 02 10:11:37 crc kubenswrapper[4934]: W1002 10:11:37.272077 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfb4baab_8e79_44ed_8d2a_7da87c069cc2.slice/crio-66b7561fb7ca1e61258bccdc813f59086e04d7cc316812de7a59869e785b855e WatchSource:0}: Error finding container 66b7561fb7ca1e61258bccdc813f59086e04d7cc316812de7a59869e785b855e: Status 404 returned error can't find the container with id 66b7561fb7ca1e61258bccdc813f59086e04d7cc316812de7a59869e785b855e Oct 02 10:11:37 crc kubenswrapper[4934]: I1002 10:11:37.718898 4934 generic.go:334] "Generic (PLEG): container finished" podID="cfb4baab-8e79-44ed-8d2a-7da87c069cc2" containerID="e217e6c9cfce0d880a3b6a436512d77b23559253caf40f90d8a6c8f6ba066fe6" exitCode=0 Oct 02 10:11:37 crc kubenswrapper[4934]: I1002 10:11:37.719002 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-77cj7" event={"ID":"cfb4baab-8e79-44ed-8d2a-7da87c069cc2","Type":"ContainerDied","Data":"e217e6c9cfce0d880a3b6a436512d77b23559253caf40f90d8a6c8f6ba066fe6"} Oct 02 10:11:37 crc kubenswrapper[4934]: I1002 10:11:37.719290 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-77cj7" event={"ID":"cfb4baab-8e79-44ed-8d2a-7da87c069cc2","Type":"ContainerStarted","Data":"66b7561fb7ca1e61258bccdc813f59086e04d7cc316812de7a59869e785b855e"} Oct 02 10:11:39 crc kubenswrapper[4934]: I1002 10:11:39.740660 4934 generic.go:334] "Generic (PLEG): container finished" podID="cfb4baab-8e79-44ed-8d2a-7da87c069cc2" containerID="d01b3e6768d9b65b1ae0046135fa852192e761bed90c80ea153260478a0c0f5e" exitCode=0 Oct 02 10:11:39 crc kubenswrapper[4934]: I1002 10:11:39.740751 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-77cj7" event={"ID":"cfb4baab-8e79-44ed-8d2a-7da87c069cc2","Type":"ContainerDied","Data":"d01b3e6768d9b65b1ae0046135fa852192e761bed90c80ea153260478a0c0f5e"} Oct 02 10:11:40 crc kubenswrapper[4934]: I1002 10:11:40.753134 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-77cj7" event={"ID":"cfb4baab-8e79-44ed-8d2a-7da87c069cc2","Type":"ContainerStarted","Data":"e8337a262ee7aad19d59cecd8c1c7a0a2b5aa57a2cde4ecdc77477461bfc5d1e"} Oct 02 10:11:40 crc kubenswrapper[4934]: I1002 10:11:40.756495 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-65wtc_1146cda2-78fc-40c5-b0d2-f8c19368fa2c/ovs-vswitchd/0.log" Oct 02 10:11:40 crc kubenswrapper[4934]: I1002 10:11:40.757630 4934 generic.go:334] "Generic (PLEG): container finished" podID="1146cda2-78fc-40c5-b0d2-f8c19368fa2c" containerID="c63d866adb61766a772229b32b348b743c5300884ef0ae4924c680c6cb85c6e0" exitCode=137 Oct 02 10:11:40 crc kubenswrapper[4934]: I1002 10:11:40.757663 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-65wtc" event={"ID":"1146cda2-78fc-40c5-b0d2-f8c19368fa2c","Type":"ContainerDied","Data":"c63d866adb61766a772229b32b348b743c5300884ef0ae4924c680c6cb85c6e0"} Oct 02 10:11:40 crc kubenswrapper[4934]: I1002 10:11:40.776755 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-77cj7" podStartSLOduration=2.319007055 podStartE2EDuration="4.776737353s" podCreationTimestamp="2025-10-02 10:11:36 +0000 UTC" firstStartedPulling="2025-10-02 10:11:37.720303397 +0000 UTC m=+1369.472944919" lastFinishedPulling="2025-10-02 10:11:40.178033695 +0000 UTC m=+1371.930675217" observedRunningTime="2025-10-02 10:11:40.774496691 +0000 UTC m=+1372.527138233" watchObservedRunningTime="2025-10-02 10:11:40.776737353 +0000 UTC m=+1372.529378875" Oct 02 10:11:40 crc kubenswrapper[4934]: I1002 10:11:40.801244 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-65wtc_1146cda2-78fc-40c5-b0d2-f8c19368fa2c/ovs-vswitchd/0.log" Oct 02 10:11:40 crc kubenswrapper[4934]: I1002 10:11:40.801908 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-65wtc" Oct 02 10:11:40 crc kubenswrapper[4934]: I1002 10:11:40.888701 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-var-run\") pod \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\" (UID: \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\") " Oct 02 10:11:40 crc kubenswrapper[4934]: I1002 10:11:40.888750 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-etc-ovs\") pod \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\" (UID: \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\") " Oct 02 10:11:40 crc kubenswrapper[4934]: I1002 10:11:40.888801 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8x84\" (UniqueName: \"kubernetes.io/projected/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-kube-api-access-x8x84\") pod \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\" (UID: \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\") " Oct 02 10:11:40 crc kubenswrapper[4934]: I1002 10:11:40.888854 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-scripts\") pod \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\" (UID: \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\") " Oct 02 10:11:40 crc kubenswrapper[4934]: I1002 10:11:40.888886 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-var-lib\") pod \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\" (UID: \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\") " Oct 02 10:11:40 crc kubenswrapper[4934]: I1002 10:11:40.888915 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-var-log\") pod \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\" (UID: \"1146cda2-78fc-40c5-b0d2-f8c19368fa2c\") " Oct 02 10:11:40 crc kubenswrapper[4934]: I1002 10:11:40.888917 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-var-run" (OuterVolumeSpecName: "var-run") pod "1146cda2-78fc-40c5-b0d2-f8c19368fa2c" (UID: "1146cda2-78fc-40c5-b0d2-f8c19368fa2c"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:11:40 crc kubenswrapper[4934]: I1002 10:11:40.889238 4934 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-var-run\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:40 crc kubenswrapper[4934]: I1002 10:11:40.889470 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "1146cda2-78fc-40c5-b0d2-f8c19368fa2c" (UID: "1146cda2-78fc-40c5-b0d2-f8c19368fa2c"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:11:40 crc kubenswrapper[4934]: I1002 10:11:40.889518 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-var-lib" (OuterVolumeSpecName: "var-lib") pod "1146cda2-78fc-40c5-b0d2-f8c19368fa2c" (UID: "1146cda2-78fc-40c5-b0d2-f8c19368fa2c"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:11:40 crc kubenswrapper[4934]: I1002 10:11:40.889869 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-var-log" (OuterVolumeSpecName: "var-log") pod "1146cda2-78fc-40c5-b0d2-f8c19368fa2c" (UID: "1146cda2-78fc-40c5-b0d2-f8c19368fa2c"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:11:40 crc kubenswrapper[4934]: I1002 10:11:40.890444 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-scripts" (OuterVolumeSpecName: "scripts") pod "1146cda2-78fc-40c5-b0d2-f8c19368fa2c" (UID: "1146cda2-78fc-40c5-b0d2-f8c19368fa2c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:11:40 crc kubenswrapper[4934]: I1002 10:11:40.895651 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-kube-api-access-x8x84" (OuterVolumeSpecName: "kube-api-access-x8x84") pod "1146cda2-78fc-40c5-b0d2-f8c19368fa2c" (UID: "1146cda2-78fc-40c5-b0d2-f8c19368fa2c"). InnerVolumeSpecName "kube-api-access-x8x84". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:40 crc kubenswrapper[4934]: I1002 10:11:40.990403 4934 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-etc-ovs\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:40 crc kubenswrapper[4934]: I1002 10:11:40.990441 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8x84\" (UniqueName: \"kubernetes.io/projected/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-kube-api-access-x8x84\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:40 crc kubenswrapper[4934]: I1002 10:11:40.990454 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:40 crc kubenswrapper[4934]: I1002 10:11:40.990464 4934 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-var-lib\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:40 crc kubenswrapper[4934]: I1002 10:11:40.990474 4934 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/1146cda2-78fc-40c5-b0d2-f8c19368fa2c-var-log\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.605725 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.702250 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a95a16-6bdd-4c08-8340-13fd85455b7a-combined-ca-bundle\") pod \"94a95a16-6bdd-4c08-8340-13fd85455b7a\" (UID: \"94a95a16-6bdd-4c08-8340-13fd85455b7a\") " Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.702363 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n69mk\" (UniqueName: \"kubernetes.io/projected/94a95a16-6bdd-4c08-8340-13fd85455b7a-kube-api-access-n69mk\") pod \"94a95a16-6bdd-4c08-8340-13fd85455b7a\" (UID: \"94a95a16-6bdd-4c08-8340-13fd85455b7a\") " Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.702398 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94a95a16-6bdd-4c08-8340-13fd85455b7a-config-data\") pod \"94a95a16-6bdd-4c08-8340-13fd85455b7a\" (UID: \"94a95a16-6bdd-4c08-8340-13fd85455b7a\") " Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.702436 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/94a95a16-6bdd-4c08-8340-13fd85455b7a-config-data-custom\") pod \"94a95a16-6bdd-4c08-8340-13fd85455b7a\" (UID: \"94a95a16-6bdd-4c08-8340-13fd85455b7a\") " Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.702496 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94a95a16-6bdd-4c08-8340-13fd85455b7a-scripts\") pod \"94a95a16-6bdd-4c08-8340-13fd85455b7a\" (UID: \"94a95a16-6bdd-4c08-8340-13fd85455b7a\") " Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.702545 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/94a95a16-6bdd-4c08-8340-13fd85455b7a-etc-machine-id\") pod \"94a95a16-6bdd-4c08-8340-13fd85455b7a\" (UID: \"94a95a16-6bdd-4c08-8340-13fd85455b7a\") " Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.705846 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/94a95a16-6bdd-4c08-8340-13fd85455b7a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "94a95a16-6bdd-4c08-8340-13fd85455b7a" (UID: "94a95a16-6bdd-4c08-8340-13fd85455b7a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.725095 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94a95a16-6bdd-4c08-8340-13fd85455b7a-kube-api-access-n69mk" (OuterVolumeSpecName: "kube-api-access-n69mk") pod "94a95a16-6bdd-4c08-8340-13fd85455b7a" (UID: "94a95a16-6bdd-4c08-8340-13fd85455b7a"). InnerVolumeSpecName "kube-api-access-n69mk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.739972 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n69mk\" (UniqueName: \"kubernetes.io/projected/94a95a16-6bdd-4c08-8340-13fd85455b7a-kube-api-access-n69mk\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.740023 4934 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/94a95a16-6bdd-4c08-8340-13fd85455b7a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.743804 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94a95a16-6bdd-4c08-8340-13fd85455b7a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "94a95a16-6bdd-4c08-8340-13fd85455b7a" (UID: "94a95a16-6bdd-4c08-8340-13fd85455b7a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.754854 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94a95a16-6bdd-4c08-8340-13fd85455b7a-scripts" (OuterVolumeSpecName: "scripts") pod "94a95a16-6bdd-4c08-8340-13fd85455b7a" (UID: "94a95a16-6bdd-4c08-8340-13fd85455b7a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.763206 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94a95a16-6bdd-4c08-8340-13fd85455b7a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "94a95a16-6bdd-4c08-8340-13fd85455b7a" (UID: "94a95a16-6bdd-4c08-8340-13fd85455b7a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.777310 4934 generic.go:334] "Generic (PLEG): container finished" podID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerID="1c65da253bd0c434afbafb0cd05ae9a507db27a1ff5ce1626446a181ca57c304" exitCode=137 Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.777372 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerDied","Data":"1c65da253bd0c434afbafb0cd05ae9a507db27a1ff5ce1626446a181ca57c304"} Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.779398 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-65wtc_1146cda2-78fc-40c5-b0d2-f8c19368fa2c/ovs-vswitchd/0.log" Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.780210 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-65wtc" event={"ID":"1146cda2-78fc-40c5-b0d2-f8c19368fa2c","Type":"ContainerDied","Data":"65bf9dacd49ccb2078166d218e8e99159a5e3b951f1f7cdf8a5023d8b0987280"} Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.780286 4934 scope.go:117] "RemoveContainer" containerID="c63d866adb61766a772229b32b348b743c5300884ef0ae4924c680c6cb85c6e0" Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.780587 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-65wtc" Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.782768 4934 generic.go:334] "Generic (PLEG): container finished" podID="94a95a16-6bdd-4c08-8340-13fd85455b7a" containerID="9c952faa06a3569c88d733912510a64a2634e55c3359c72da98dba68500b8acf" exitCode=137 Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.783963 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"94a95a16-6bdd-4c08-8340-13fd85455b7a","Type":"ContainerDied","Data":"9c952faa06a3569c88d733912510a64a2634e55c3359c72da98dba68500b8acf"} Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.784044 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"94a95a16-6bdd-4c08-8340-13fd85455b7a","Type":"ContainerDied","Data":"9998933aed61ac0b0458b31757054f9862222bb41ad966c21139956e6a6eb297"} Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.785422 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.816758 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94a95a16-6bdd-4c08-8340-13fd85455b7a-config-data" (OuterVolumeSpecName: "config-data") pod "94a95a16-6bdd-4c08-8340-13fd85455b7a" (UID: "94a95a16-6bdd-4c08-8340-13fd85455b7a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.840857 4934 scope.go:117] "RemoveContainer" containerID="9f1bc28ee655a461d4d083b7fa1dad3e8809d4253e14fa90ef0107c3618e2892" Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.840938 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94a95a16-6bdd-4c08-8340-13fd85455b7a-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.840962 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a95a16-6bdd-4c08-8340-13fd85455b7a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.840974 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94a95a16-6bdd-4c08-8340-13fd85455b7a-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.840984 4934 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/94a95a16-6bdd-4c08-8340-13fd85455b7a-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.841066 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-65wtc"] Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.850873 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-65wtc"] Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.859404 4934 scope.go:117] "RemoveContainer" containerID="73698957b5b31581f4624457e2c99dc25e8485557f6ad04a08edcb05551c73f3" Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.885725 4934 scope.go:117] "RemoveContainer" containerID="7f805c812aa2a019f6996c4840760eec127d8f206fe82dbfef04db5d8dadd02c" Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.894083 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.907712 4934 scope.go:117] "RemoveContainer" containerID="9c952faa06a3569c88d733912510a64a2634e55c3359c72da98dba68500b8acf" Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.928435 4934 scope.go:117] "RemoveContainer" containerID="7f805c812aa2a019f6996c4840760eec127d8f206fe82dbfef04db5d8dadd02c" Oct 02 10:11:41 crc kubenswrapper[4934]: E1002 10:11:41.928952 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f805c812aa2a019f6996c4840760eec127d8f206fe82dbfef04db5d8dadd02c\": container with ID starting with 7f805c812aa2a019f6996c4840760eec127d8f206fe82dbfef04db5d8dadd02c not found: ID does not exist" containerID="7f805c812aa2a019f6996c4840760eec127d8f206fe82dbfef04db5d8dadd02c" Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.928989 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f805c812aa2a019f6996c4840760eec127d8f206fe82dbfef04db5d8dadd02c"} err="failed to get container status \"7f805c812aa2a019f6996c4840760eec127d8f206fe82dbfef04db5d8dadd02c\": rpc error: code = NotFound desc = could not find container \"7f805c812aa2a019f6996c4840760eec127d8f206fe82dbfef04db5d8dadd02c\": container with ID starting with 7f805c812aa2a019f6996c4840760eec127d8f206fe82dbfef04db5d8dadd02c not found: ID does not exist" Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.929015 4934 scope.go:117] "RemoveContainer" containerID="9c952faa06a3569c88d733912510a64a2634e55c3359c72da98dba68500b8acf" Oct 02 10:11:41 crc kubenswrapper[4934]: E1002 10:11:41.929307 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c952faa06a3569c88d733912510a64a2634e55c3359c72da98dba68500b8acf\": container with ID starting with 9c952faa06a3569c88d733912510a64a2634e55c3359c72da98dba68500b8acf not found: ID does not exist" containerID="9c952faa06a3569c88d733912510a64a2634e55c3359c72da98dba68500b8acf" Oct 02 10:11:41 crc kubenswrapper[4934]: I1002 10:11:41.929363 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c952faa06a3569c88d733912510a64a2634e55c3359c72da98dba68500b8acf"} err="failed to get container status \"9c952faa06a3569c88d733912510a64a2634e55c3359c72da98dba68500b8acf\": rpc error: code = NotFound desc = could not find container \"9c952faa06a3569c88d733912510a64a2634e55c3359c72da98dba68500b8acf\": container with ID starting with 9c952faa06a3569c88d733912510a64a2634e55c3359c72da98dba68500b8acf not found: ID does not exist" Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.045207 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8cebd152-5aaf-4543-be11-7ba327cfdf0e-etc-swift\") pod \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\" (UID: \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\") " Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.045267 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/8cebd152-5aaf-4543-be11-7ba327cfdf0e-cache\") pod \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\" (UID: \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\") " Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.045284 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/8cebd152-5aaf-4543-be11-7ba327cfdf0e-lock\") pod \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\" (UID: \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\") " Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.045340 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\" (UID: \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\") " Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.045407 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zt2qp\" (UniqueName: \"kubernetes.io/projected/8cebd152-5aaf-4543-be11-7ba327cfdf0e-kube-api-access-zt2qp\") pod \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\" (UID: \"8cebd152-5aaf-4543-be11-7ba327cfdf0e\") " Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.046291 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cebd152-5aaf-4543-be11-7ba327cfdf0e-lock" (OuterVolumeSpecName: "lock") pod "8cebd152-5aaf-4543-be11-7ba327cfdf0e" (UID: "8cebd152-5aaf-4543-be11-7ba327cfdf0e"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.046697 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cebd152-5aaf-4543-be11-7ba327cfdf0e-cache" (OuterVolumeSpecName: "cache") pod "8cebd152-5aaf-4543-be11-7ba327cfdf0e" (UID: "8cebd152-5aaf-4543-be11-7ba327cfdf0e"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.048548 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "swift") pod "8cebd152-5aaf-4543-be11-7ba327cfdf0e" (UID: "8cebd152-5aaf-4543-be11-7ba327cfdf0e"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.048750 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cebd152-5aaf-4543-be11-7ba327cfdf0e-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "8cebd152-5aaf-4543-be11-7ba327cfdf0e" (UID: "8cebd152-5aaf-4543-be11-7ba327cfdf0e"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.049027 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cebd152-5aaf-4543-be11-7ba327cfdf0e-kube-api-access-zt2qp" (OuterVolumeSpecName: "kube-api-access-zt2qp") pod "8cebd152-5aaf-4543-be11-7ba327cfdf0e" (UID: "8cebd152-5aaf-4543-be11-7ba327cfdf0e"). InnerVolumeSpecName "kube-api-access-zt2qp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.118996 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.127989 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.147055 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zt2qp\" (UniqueName: \"kubernetes.io/projected/8cebd152-5aaf-4543-be11-7ba327cfdf0e-kube-api-access-zt2qp\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.147085 4934 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8cebd152-5aaf-4543-be11-7ba327cfdf0e-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.147093 4934 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/8cebd152-5aaf-4543-be11-7ba327cfdf0e-cache\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.147103 4934 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/8cebd152-5aaf-4543-be11-7ba327cfdf0e-lock\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.147134 4934 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.162884 4934 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.248467 4934 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.801632 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.801621 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8cebd152-5aaf-4543-be11-7ba327cfdf0e","Type":"ContainerDied","Data":"44025b72bcf24f1079aba739bc296bb9f110537bfa5d3c6ec0f9d48d0c6b4a07"} Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.801853 4934 scope.go:117] "RemoveContainer" containerID="1c65da253bd0c434afbafb0cd05ae9a507db27a1ff5ce1626446a181ca57c304" Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.840893 4934 scope.go:117] "RemoveContainer" containerID="9373e5fbc3b166030467efc3d48ed6860fef2654fa4d21f813440b32765b3063" Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.849324 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.870903 4934 scope.go:117] "RemoveContainer" containerID="2d1cebd6bce0741d006ac48e38db4c872771c02036472a9fa71980ddfb72cef5" Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.871071 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.890783 4934 scope.go:117] "RemoveContainer" containerID="6b7e46bf6361363ea3ce0effc8f79e7c186a4d64e12c8053c444d0f89313722e" Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.907980 4934 scope.go:117] "RemoveContainer" containerID="169e61c78b12ac121ef1306849a0fe5e05cc7125ba6f6a238a652aac3445bad8" Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.945483 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1146cda2-78fc-40c5-b0d2-f8c19368fa2c" path="/var/lib/kubelet/pods/1146cda2-78fc-40c5-b0d2-f8c19368fa2c/volumes" Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.947049 4934 scope.go:117] "RemoveContainer" containerID="d6fc4c846f604ef5014de443285ee7dab569b41691babd5a6981585d8007df21" Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.947228 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" path="/var/lib/kubelet/pods/8cebd152-5aaf-4543-be11-7ba327cfdf0e/volumes" Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.950263 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94a95a16-6bdd-4c08-8340-13fd85455b7a" path="/var/lib/kubelet/pods/94a95a16-6bdd-4c08-8340-13fd85455b7a/volumes" Oct 02 10:11:42 crc kubenswrapper[4934]: E1002 10:11:42.953619 4934 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cebd152_5aaf_4543_be11_7ba327cfdf0e.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cebd152_5aaf_4543_be11_7ba327cfdf0e.slice/crio-44025b72bcf24f1079aba739bc296bb9f110537bfa5d3c6ec0f9d48d0c6b4a07\": RecentStats: unable to find data in memory cache]" Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.968936 4934 scope.go:117] "RemoveContainer" containerID="1e2427a94989f9a77ca59046a739fc7849881116c29e79c766ddc52debf75ec3" Oct 02 10:11:42 crc kubenswrapper[4934]: I1002 10:11:42.989543 4934 scope.go:117] "RemoveContainer" containerID="883824276a85ba5f915880ccf065542294a5a3959ead2ee9e02e556f4e13a641" Oct 02 10:11:43 crc kubenswrapper[4934]: I1002 10:11:43.009230 4934 scope.go:117] "RemoveContainer" containerID="022023c76cc2ba5ec65353c6ec2387833901f691565a07dc55b9a6a09bf9871c" Oct 02 10:11:43 crc kubenswrapper[4934]: I1002 10:11:43.027710 4934 scope.go:117] "RemoveContainer" containerID="da203f48b1e4a0c44c371939bd0a1b6ee7159680a2a3fac865d9476c3b7b9c05" Oct 02 10:11:43 crc kubenswrapper[4934]: I1002 10:11:43.044597 4934 scope.go:117] "RemoveContainer" containerID="c0f5815d02eff4beefc969cb8468e9b85a60d871f99cc29f33aa27f2cd50039f" Oct 02 10:11:43 crc kubenswrapper[4934]: I1002 10:11:43.066632 4934 scope.go:117] "RemoveContainer" containerID="65c0b15d452f5559c24817138f1d42f00b5d08d34eff5f6a036df1ded5c70ccb" Oct 02 10:11:43 crc kubenswrapper[4934]: I1002 10:11:43.086938 4934 scope.go:117] "RemoveContainer" containerID="5c785fb317467101f9ff407243fb85596511c5e525db3832b38df04ed00fec73" Oct 02 10:11:43 crc kubenswrapper[4934]: I1002 10:11:43.105503 4934 scope.go:117] "RemoveContainer" containerID="283ced1ad0dcd3181290d000eab3ac44a3033b8e32fa0ca0cd8ef03d57a7deee" Oct 02 10:11:43 crc kubenswrapper[4934]: I1002 10:11:43.128077 4934 scope.go:117] "RemoveContainer" containerID="478b442a459773165d46dac69a02032f32c9b37d6398eff28160ced5dd0ee4b4" Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.497675 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement5c1f-account-delete-2jjrd" Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.504140 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance82a2-account-delete-sv9vf" Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.518824 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder36fe-account-delete-n2bk5" Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.595115 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tqdj\" (UniqueName: \"kubernetes.io/projected/8679bb63-5c56-42a3-9b2c-42303e3afef6-kube-api-access-8tqdj\") pod \"8679bb63-5c56-42a3-9b2c-42303e3afef6\" (UID: \"8679bb63-5c56-42a3-9b2c-42303e3afef6\") " Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.595209 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44tbd\" (UniqueName: \"kubernetes.io/projected/5ceb70f8-5cc3-4e91-9bc1-6fd0b7575efc-kube-api-access-44tbd\") pod \"5ceb70f8-5cc3-4e91-9bc1-6fd0b7575efc\" (UID: \"5ceb70f8-5cc3-4e91-9bc1-6fd0b7575efc\") " Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.595324 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hj7sq\" (UniqueName: \"kubernetes.io/projected/bf007c3a-2ba5-4d94-a789-a5413787d6d8-kube-api-access-hj7sq\") pod \"bf007c3a-2ba5-4d94-a789-a5413787d6d8\" (UID: \"bf007c3a-2ba5-4d94-a789-a5413787d6d8\") " Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.600719 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf007c3a-2ba5-4d94-a789-a5413787d6d8-kube-api-access-hj7sq" (OuterVolumeSpecName: "kube-api-access-hj7sq") pod "bf007c3a-2ba5-4d94-a789-a5413787d6d8" (UID: "bf007c3a-2ba5-4d94-a789-a5413787d6d8"). InnerVolumeSpecName "kube-api-access-hj7sq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.601655 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8679bb63-5c56-42a3-9b2c-42303e3afef6-kube-api-access-8tqdj" (OuterVolumeSpecName: "kube-api-access-8tqdj") pod "8679bb63-5c56-42a3-9b2c-42303e3afef6" (UID: "8679bb63-5c56-42a3-9b2c-42303e3afef6"). InnerVolumeSpecName "kube-api-access-8tqdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.602044 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ceb70f8-5cc3-4e91-9bc1-6fd0b7575efc-kube-api-access-44tbd" (OuterVolumeSpecName: "kube-api-access-44tbd") pod "5ceb70f8-5cc3-4e91-9bc1-6fd0b7575efc" (UID: "5ceb70f8-5cc3-4e91-9bc1-6fd0b7575efc"). InnerVolumeSpecName "kube-api-access-44tbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.696571 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hj7sq\" (UniqueName: \"kubernetes.io/projected/bf007c3a-2ba5-4d94-a789-a5413787d6d8-kube-api-access-hj7sq\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.696617 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tqdj\" (UniqueName: \"kubernetes.io/projected/8679bb63-5c56-42a3-9b2c-42303e3afef6-kube-api-access-8tqdj\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.696626 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44tbd\" (UniqueName: \"kubernetes.io/projected/5ceb70f8-5cc3-4e91-9bc1-6fd0b7575efc-kube-api-access-44tbd\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.841735 4934 generic.go:334] "Generic (PLEG): container finished" podID="5ceb70f8-5cc3-4e91-9bc1-6fd0b7575efc" containerID="3bc9955bd39440cdf80a78378bd0626b6c3f71633a26259bbf572670b66ff302" exitCode=137 Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.841804 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement5c1f-account-delete-2jjrd" event={"ID":"5ceb70f8-5cc3-4e91-9bc1-6fd0b7575efc","Type":"ContainerDied","Data":"3bc9955bd39440cdf80a78378bd0626b6c3f71633a26259bbf572670b66ff302"} Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.841819 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement5c1f-account-delete-2jjrd" Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.841844 4934 scope.go:117] "RemoveContainer" containerID="3bc9955bd39440cdf80a78378bd0626b6c3f71633a26259bbf572670b66ff302" Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.841832 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement5c1f-account-delete-2jjrd" event={"ID":"5ceb70f8-5cc3-4e91-9bc1-6fd0b7575efc","Type":"ContainerDied","Data":"7ed9155cd1ead1214836ae38ef467cb61192842e0c3254a04b353ba0719115b2"} Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.846373 4934 generic.go:334] "Generic (PLEG): container finished" podID="bf007c3a-2ba5-4d94-a789-a5413787d6d8" containerID="114be7966ff7fb8bfc8736837ac939a0abe593bac1ebf6b0afb5973b3356ee6d" exitCode=137 Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.846441 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder36fe-account-delete-n2bk5" Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.846458 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder36fe-account-delete-n2bk5" event={"ID":"bf007c3a-2ba5-4d94-a789-a5413787d6d8","Type":"ContainerDied","Data":"114be7966ff7fb8bfc8736837ac939a0abe593bac1ebf6b0afb5973b3356ee6d"} Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.846500 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder36fe-account-delete-n2bk5" event={"ID":"bf007c3a-2ba5-4d94-a789-a5413787d6d8","Type":"ContainerDied","Data":"343e2c45aadc061e539aead3e470e91d2b221db608fa86bf327bb7d852fe8a92"} Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.848823 4934 generic.go:334] "Generic (PLEG): container finished" podID="8679bb63-5c56-42a3-9b2c-42303e3afef6" containerID="2696284bd53dcc40f609d9c2ea7a8c80534e0456354a1ab9817e35d0d0a4d9a4" exitCode=137 Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.848868 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance82a2-account-delete-sv9vf" event={"ID":"8679bb63-5c56-42a3-9b2c-42303e3afef6","Type":"ContainerDied","Data":"2696284bd53dcc40f609d9c2ea7a8c80534e0456354a1ab9817e35d0d0a4d9a4"} Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.848890 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance82a2-account-delete-sv9vf" event={"ID":"8679bb63-5c56-42a3-9b2c-42303e3afef6","Type":"ContainerDied","Data":"d6b82a35f04bd2f56ef75ad50a445e0346ac7b81813f75928def62ce66c8ec56"} Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.848962 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance82a2-account-delete-sv9vf" Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.887118 4934 scope.go:117] "RemoveContainer" containerID="3bc9955bd39440cdf80a78378bd0626b6c3f71633a26259bbf572670b66ff302" Oct 02 10:11:45 crc kubenswrapper[4934]: E1002 10:11:45.887806 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bc9955bd39440cdf80a78378bd0626b6c3f71633a26259bbf572670b66ff302\": container with ID starting with 3bc9955bd39440cdf80a78378bd0626b6c3f71633a26259bbf572670b66ff302 not found: ID does not exist" containerID="3bc9955bd39440cdf80a78378bd0626b6c3f71633a26259bbf572670b66ff302" Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.887836 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bc9955bd39440cdf80a78378bd0626b6c3f71633a26259bbf572670b66ff302"} err="failed to get container status \"3bc9955bd39440cdf80a78378bd0626b6c3f71633a26259bbf572670b66ff302\": rpc error: code = NotFound desc = could not find container \"3bc9955bd39440cdf80a78378bd0626b6c3f71633a26259bbf572670b66ff302\": container with ID starting with 3bc9955bd39440cdf80a78378bd0626b6c3f71633a26259bbf572670b66ff302 not found: ID does not exist" Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.887852 4934 scope.go:117] "RemoveContainer" containerID="114be7966ff7fb8bfc8736837ac939a0abe593bac1ebf6b0afb5973b3356ee6d" Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.903570 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance82a2-account-delete-sv9vf"] Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.913513 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance82a2-account-delete-sv9vf"] Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.919251 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement5c1f-account-delete-2jjrd"] Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.926433 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement5c1f-account-delete-2jjrd"] Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.930881 4934 scope.go:117] "RemoveContainer" containerID="114be7966ff7fb8bfc8736837ac939a0abe593bac1ebf6b0afb5973b3356ee6d" Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.932208 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder36fe-account-delete-n2bk5"] Oct 02 10:11:45 crc kubenswrapper[4934]: E1002 10:11:45.932433 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"114be7966ff7fb8bfc8736837ac939a0abe593bac1ebf6b0afb5973b3356ee6d\": container with ID starting with 114be7966ff7fb8bfc8736837ac939a0abe593bac1ebf6b0afb5973b3356ee6d not found: ID does not exist" containerID="114be7966ff7fb8bfc8736837ac939a0abe593bac1ebf6b0afb5973b3356ee6d" Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.932469 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"114be7966ff7fb8bfc8736837ac939a0abe593bac1ebf6b0afb5973b3356ee6d"} err="failed to get container status \"114be7966ff7fb8bfc8736837ac939a0abe593bac1ebf6b0afb5973b3356ee6d\": rpc error: code = NotFound desc = could not find container \"114be7966ff7fb8bfc8736837ac939a0abe593bac1ebf6b0afb5973b3356ee6d\": container with ID starting with 114be7966ff7fb8bfc8736837ac939a0abe593bac1ebf6b0afb5973b3356ee6d not found: ID does not exist" Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.932498 4934 scope.go:117] "RemoveContainer" containerID="2696284bd53dcc40f609d9c2ea7a8c80534e0456354a1ab9817e35d0d0a4d9a4" Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.937233 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder36fe-account-delete-n2bk5"] Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.957803 4934 scope.go:117] "RemoveContainer" containerID="2696284bd53dcc40f609d9c2ea7a8c80534e0456354a1ab9817e35d0d0a4d9a4" Oct 02 10:11:45 crc kubenswrapper[4934]: E1002 10:11:45.958222 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2696284bd53dcc40f609d9c2ea7a8c80534e0456354a1ab9817e35d0d0a4d9a4\": container with ID starting with 2696284bd53dcc40f609d9c2ea7a8c80534e0456354a1ab9817e35d0d0a4d9a4 not found: ID does not exist" containerID="2696284bd53dcc40f609d9c2ea7a8c80534e0456354a1ab9817e35d0d0a4d9a4" Oct 02 10:11:45 crc kubenswrapper[4934]: I1002 10:11:45.958256 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2696284bd53dcc40f609d9c2ea7a8c80534e0456354a1ab9817e35d0d0a4d9a4"} err="failed to get container status \"2696284bd53dcc40f609d9c2ea7a8c80534e0456354a1ab9817e35d0d0a4d9a4\": rpc error: code = NotFound desc = could not find container \"2696284bd53dcc40f609d9c2ea7a8c80534e0456354a1ab9817e35d0d0a4d9a4\": container with ID starting with 2696284bd53dcc40f609d9c2ea7a8c80534e0456354a1ab9817e35d0d0a4d9a4 not found: ID does not exist" Oct 02 10:11:46 crc kubenswrapper[4934]: I1002 10:11:46.829912 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-77cj7" Oct 02 10:11:46 crc kubenswrapper[4934]: I1002 10:11:46.829984 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-77cj7" Oct 02 10:11:46 crc kubenswrapper[4934]: I1002 10:11:46.883929 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-77cj7" Oct 02 10:11:46 crc kubenswrapper[4934]: I1002 10:11:46.925637 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ceb70f8-5cc3-4e91-9bc1-6fd0b7575efc" path="/var/lib/kubelet/pods/5ceb70f8-5cc3-4e91-9bc1-6fd0b7575efc/volumes" Oct 02 10:11:46 crc kubenswrapper[4934]: I1002 10:11:46.926796 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8679bb63-5c56-42a3-9b2c-42303e3afef6" path="/var/lib/kubelet/pods/8679bb63-5c56-42a3-9b2c-42303e3afef6/volumes" Oct 02 10:11:46 crc kubenswrapper[4934]: I1002 10:11:46.927914 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf007c3a-2ba5-4d94-a789-a5413787d6d8" path="/var/lib/kubelet/pods/bf007c3a-2ba5-4d94-a789-a5413787d6d8/volumes" Oct 02 10:11:46 crc kubenswrapper[4934]: I1002 10:11:46.949295 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-77cj7" Oct 02 10:11:47 crc kubenswrapper[4934]: I1002 10:11:47.116744 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-77cj7"] Oct 02 10:11:47 crc kubenswrapper[4934]: I1002 10:11:47.728950 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi26a6-account-delete-ng2mb" Oct 02 10:11:47 crc kubenswrapper[4934]: I1002 10:11:47.829410 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlc5h\" (UniqueName: \"kubernetes.io/projected/a7f7a4d6-c246-4c88-85e1-573fd07ec33f-kube-api-access-jlc5h\") pod \"a7f7a4d6-c246-4c88-85e1-573fd07ec33f\" (UID: \"a7f7a4d6-c246-4c88-85e1-573fd07ec33f\") " Oct 02 10:11:47 crc kubenswrapper[4934]: I1002 10:11:47.835703 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7f7a4d6-c246-4c88-85e1-573fd07ec33f-kube-api-access-jlc5h" (OuterVolumeSpecName: "kube-api-access-jlc5h") pod "a7f7a4d6-c246-4c88-85e1-573fd07ec33f" (UID: "a7f7a4d6-c246-4c88-85e1-573fd07ec33f"). InnerVolumeSpecName "kube-api-access-jlc5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:47 crc kubenswrapper[4934]: I1002 10:11:47.874166 4934 generic.go:334] "Generic (PLEG): container finished" podID="a7f7a4d6-c246-4c88-85e1-573fd07ec33f" containerID="4251b40d65b71246a67772b89828963553528ad5d1e49e7473e8cac466296c4b" exitCode=137 Oct 02 10:11:47 crc kubenswrapper[4934]: I1002 10:11:47.874269 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi26a6-account-delete-ng2mb" Oct 02 10:11:47 crc kubenswrapper[4934]: I1002 10:11:47.874338 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi26a6-account-delete-ng2mb" event={"ID":"a7f7a4d6-c246-4c88-85e1-573fd07ec33f","Type":"ContainerDied","Data":"4251b40d65b71246a67772b89828963553528ad5d1e49e7473e8cac466296c4b"} Oct 02 10:11:47 crc kubenswrapper[4934]: I1002 10:11:47.874382 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi26a6-account-delete-ng2mb" event={"ID":"a7f7a4d6-c246-4c88-85e1-573fd07ec33f","Type":"ContainerDied","Data":"bc902f497027122db034b741d4e334263ead76a6fc25aa6afc21ca8303950f02"} Oct 02 10:11:47 crc kubenswrapper[4934]: I1002 10:11:47.874412 4934 scope.go:117] "RemoveContainer" containerID="4251b40d65b71246a67772b89828963553528ad5d1e49e7473e8cac466296c4b" Oct 02 10:11:47 crc kubenswrapper[4934]: I1002 10:11:47.905223 4934 scope.go:117] "RemoveContainer" containerID="4251b40d65b71246a67772b89828963553528ad5d1e49e7473e8cac466296c4b" Oct 02 10:11:47 crc kubenswrapper[4934]: E1002 10:11:47.906180 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4251b40d65b71246a67772b89828963553528ad5d1e49e7473e8cac466296c4b\": container with ID starting with 4251b40d65b71246a67772b89828963553528ad5d1e49e7473e8cac466296c4b not found: ID does not exist" containerID="4251b40d65b71246a67772b89828963553528ad5d1e49e7473e8cac466296c4b" Oct 02 10:11:47 crc kubenswrapper[4934]: I1002 10:11:47.906264 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4251b40d65b71246a67772b89828963553528ad5d1e49e7473e8cac466296c4b"} err="failed to get container status \"4251b40d65b71246a67772b89828963553528ad5d1e49e7473e8cac466296c4b\": rpc error: code = NotFound desc = could not find container \"4251b40d65b71246a67772b89828963553528ad5d1e49e7473e8cac466296c4b\": container with ID starting with 4251b40d65b71246a67772b89828963553528ad5d1e49e7473e8cac466296c4b not found: ID does not exist" Oct 02 10:11:47 crc kubenswrapper[4934]: I1002 10:11:47.919091 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi26a6-account-delete-ng2mb"] Oct 02 10:11:47 crc kubenswrapper[4934]: I1002 10:11:47.924252 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapi26a6-account-delete-ng2mb"] Oct 02 10:11:47 crc kubenswrapper[4934]: I1002 10:11:47.931153 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlc5h\" (UniqueName: \"kubernetes.io/projected/a7f7a4d6-c246-4c88-85e1-573fd07ec33f-kube-api-access-jlc5h\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:48 crc kubenswrapper[4934]: I1002 10:11:48.888994 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-77cj7" podUID="cfb4baab-8e79-44ed-8d2a-7da87c069cc2" containerName="registry-server" containerID="cri-o://e8337a262ee7aad19d59cecd8c1c7a0a2b5aa57a2cde4ecdc77477461bfc5d1e" gracePeriod=2 Oct 02 10:11:48 crc kubenswrapper[4934]: I1002 10:11:48.950318 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7f7a4d6-c246-4c88-85e1-573fd07ec33f" path="/var/lib/kubelet/pods/a7f7a4d6-c246-4c88-85e1-573fd07ec33f/volumes" Oct 02 10:11:49 crc kubenswrapper[4934]: I1002 10:11:49.347487 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-77cj7" Oct 02 10:11:49 crc kubenswrapper[4934]: I1002 10:11:49.464470 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfb4baab-8e79-44ed-8d2a-7da87c069cc2-utilities\") pod \"cfb4baab-8e79-44ed-8d2a-7da87c069cc2\" (UID: \"cfb4baab-8e79-44ed-8d2a-7da87c069cc2\") " Oct 02 10:11:49 crc kubenswrapper[4934]: I1002 10:11:49.464549 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffkfq\" (UniqueName: \"kubernetes.io/projected/cfb4baab-8e79-44ed-8d2a-7da87c069cc2-kube-api-access-ffkfq\") pod \"cfb4baab-8e79-44ed-8d2a-7da87c069cc2\" (UID: \"cfb4baab-8e79-44ed-8d2a-7da87c069cc2\") " Oct 02 10:11:49 crc kubenswrapper[4934]: I1002 10:11:49.464568 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfb4baab-8e79-44ed-8d2a-7da87c069cc2-catalog-content\") pod \"cfb4baab-8e79-44ed-8d2a-7da87c069cc2\" (UID: \"cfb4baab-8e79-44ed-8d2a-7da87c069cc2\") " Oct 02 10:11:49 crc kubenswrapper[4934]: I1002 10:11:49.465793 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfb4baab-8e79-44ed-8d2a-7da87c069cc2-utilities" (OuterVolumeSpecName: "utilities") pod "cfb4baab-8e79-44ed-8d2a-7da87c069cc2" (UID: "cfb4baab-8e79-44ed-8d2a-7da87c069cc2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:49 crc kubenswrapper[4934]: I1002 10:11:49.489181 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfb4baab-8e79-44ed-8d2a-7da87c069cc2-kube-api-access-ffkfq" (OuterVolumeSpecName: "kube-api-access-ffkfq") pod "cfb4baab-8e79-44ed-8d2a-7da87c069cc2" (UID: "cfb4baab-8e79-44ed-8d2a-7da87c069cc2"). InnerVolumeSpecName "kube-api-access-ffkfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:11:49 crc kubenswrapper[4934]: I1002 10:11:49.566767 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfb4baab-8e79-44ed-8d2a-7da87c069cc2-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:49 crc kubenswrapper[4934]: I1002 10:11:49.566807 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffkfq\" (UniqueName: \"kubernetes.io/projected/cfb4baab-8e79-44ed-8d2a-7da87c069cc2-kube-api-access-ffkfq\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:49 crc kubenswrapper[4934]: I1002 10:11:49.567775 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfb4baab-8e79-44ed-8d2a-7da87c069cc2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cfb4baab-8e79-44ed-8d2a-7da87c069cc2" (UID: "cfb4baab-8e79-44ed-8d2a-7da87c069cc2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:11:49 crc kubenswrapper[4934]: I1002 10:11:49.667730 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfb4baab-8e79-44ed-8d2a-7da87c069cc2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:11:49 crc kubenswrapper[4934]: I1002 10:11:49.903194 4934 generic.go:334] "Generic (PLEG): container finished" podID="cfb4baab-8e79-44ed-8d2a-7da87c069cc2" containerID="e8337a262ee7aad19d59cecd8c1c7a0a2b5aa57a2cde4ecdc77477461bfc5d1e" exitCode=0 Oct 02 10:11:49 crc kubenswrapper[4934]: I1002 10:11:49.903237 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-77cj7" event={"ID":"cfb4baab-8e79-44ed-8d2a-7da87c069cc2","Type":"ContainerDied","Data":"e8337a262ee7aad19d59cecd8c1c7a0a2b5aa57a2cde4ecdc77477461bfc5d1e"} Oct 02 10:11:49 crc kubenswrapper[4934]: I1002 10:11:49.903262 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-77cj7" event={"ID":"cfb4baab-8e79-44ed-8d2a-7da87c069cc2","Type":"ContainerDied","Data":"66b7561fb7ca1e61258bccdc813f59086e04d7cc316812de7a59869e785b855e"} Oct 02 10:11:49 crc kubenswrapper[4934]: I1002 10:11:49.903279 4934 scope.go:117] "RemoveContainer" containerID="e8337a262ee7aad19d59cecd8c1c7a0a2b5aa57a2cde4ecdc77477461bfc5d1e" Oct 02 10:11:49 crc kubenswrapper[4934]: I1002 10:11:49.903289 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-77cj7" Oct 02 10:11:49 crc kubenswrapper[4934]: I1002 10:11:49.932415 4934 scope.go:117] "RemoveContainer" containerID="d01b3e6768d9b65b1ae0046135fa852192e761bed90c80ea153260478a0c0f5e" Oct 02 10:11:49 crc kubenswrapper[4934]: I1002 10:11:49.952838 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-77cj7"] Oct 02 10:11:49 crc kubenswrapper[4934]: I1002 10:11:49.961620 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-77cj7"] Oct 02 10:11:49 crc kubenswrapper[4934]: I1002 10:11:49.976524 4934 scope.go:117] "RemoveContainer" containerID="e217e6c9cfce0d880a3b6a436512d77b23559253caf40f90d8a6c8f6ba066fe6" Oct 02 10:11:50 crc kubenswrapper[4934]: I1002 10:11:50.011129 4934 scope.go:117] "RemoveContainer" containerID="e8337a262ee7aad19d59cecd8c1c7a0a2b5aa57a2cde4ecdc77477461bfc5d1e" Oct 02 10:11:50 crc kubenswrapper[4934]: E1002 10:11:50.011776 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8337a262ee7aad19d59cecd8c1c7a0a2b5aa57a2cde4ecdc77477461bfc5d1e\": container with ID starting with e8337a262ee7aad19d59cecd8c1c7a0a2b5aa57a2cde4ecdc77477461bfc5d1e not found: ID does not exist" containerID="e8337a262ee7aad19d59cecd8c1c7a0a2b5aa57a2cde4ecdc77477461bfc5d1e" Oct 02 10:11:50 crc kubenswrapper[4934]: I1002 10:11:50.011863 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8337a262ee7aad19d59cecd8c1c7a0a2b5aa57a2cde4ecdc77477461bfc5d1e"} err="failed to get container status \"e8337a262ee7aad19d59cecd8c1c7a0a2b5aa57a2cde4ecdc77477461bfc5d1e\": rpc error: code = NotFound desc = could not find container \"e8337a262ee7aad19d59cecd8c1c7a0a2b5aa57a2cde4ecdc77477461bfc5d1e\": container with ID starting with e8337a262ee7aad19d59cecd8c1c7a0a2b5aa57a2cde4ecdc77477461bfc5d1e not found: ID does not exist" Oct 02 10:11:50 crc kubenswrapper[4934]: I1002 10:11:50.011917 4934 scope.go:117] "RemoveContainer" containerID="d01b3e6768d9b65b1ae0046135fa852192e761bed90c80ea153260478a0c0f5e" Oct 02 10:11:50 crc kubenswrapper[4934]: E1002 10:11:50.012469 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d01b3e6768d9b65b1ae0046135fa852192e761bed90c80ea153260478a0c0f5e\": container with ID starting with d01b3e6768d9b65b1ae0046135fa852192e761bed90c80ea153260478a0c0f5e not found: ID does not exist" containerID="d01b3e6768d9b65b1ae0046135fa852192e761bed90c80ea153260478a0c0f5e" Oct 02 10:11:50 crc kubenswrapper[4934]: I1002 10:11:50.012603 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d01b3e6768d9b65b1ae0046135fa852192e761bed90c80ea153260478a0c0f5e"} err="failed to get container status \"d01b3e6768d9b65b1ae0046135fa852192e761bed90c80ea153260478a0c0f5e\": rpc error: code = NotFound desc = could not find container \"d01b3e6768d9b65b1ae0046135fa852192e761bed90c80ea153260478a0c0f5e\": container with ID starting with d01b3e6768d9b65b1ae0046135fa852192e761bed90c80ea153260478a0c0f5e not found: ID does not exist" Oct 02 10:11:50 crc kubenswrapper[4934]: I1002 10:11:50.012699 4934 scope.go:117] "RemoveContainer" containerID="e217e6c9cfce0d880a3b6a436512d77b23559253caf40f90d8a6c8f6ba066fe6" Oct 02 10:11:50 crc kubenswrapper[4934]: E1002 10:11:50.013069 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e217e6c9cfce0d880a3b6a436512d77b23559253caf40f90d8a6c8f6ba066fe6\": container with ID starting with e217e6c9cfce0d880a3b6a436512d77b23559253caf40f90d8a6c8f6ba066fe6 not found: ID does not exist" containerID="e217e6c9cfce0d880a3b6a436512d77b23559253caf40f90d8a6c8f6ba066fe6" Oct 02 10:11:50 crc kubenswrapper[4934]: I1002 10:11:50.013126 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e217e6c9cfce0d880a3b6a436512d77b23559253caf40f90d8a6c8f6ba066fe6"} err="failed to get container status \"e217e6c9cfce0d880a3b6a436512d77b23559253caf40f90d8a6c8f6ba066fe6\": rpc error: code = NotFound desc = could not find container \"e217e6c9cfce0d880a3b6a436512d77b23559253caf40f90d8a6c8f6ba066fe6\": container with ID starting with e217e6c9cfce0d880a3b6a436512d77b23559253caf40f90d8a6c8f6ba066fe6 not found: ID does not exist" Oct 02 10:11:50 crc kubenswrapper[4934]: I1002 10:11:50.931870 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfb4baab-8e79-44ed-8d2a-7da87c069cc2" path="/var/lib/kubelet/pods/cfb4baab-8e79-44ed-8d2a-7da87c069cc2/volumes" Oct 02 10:11:53 crc kubenswrapper[4934]: E1002 10:11:53.173617 4934 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfb4baab_8e79_44ed_8d2a_7da87c069cc2.slice/crio-conmon-e8337a262ee7aad19d59cecd8c1c7a0a2b5aa57a2cde4ecdc77477461bfc5d1e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfb4baab_8e79_44ed_8d2a_7da87c069cc2.slice/crio-e8337a262ee7aad19d59cecd8c1c7a0a2b5aa57a2cde4ecdc77477461bfc5d1e.scope\": RecentStats: unable to find data in memory cache]" Oct 02 10:12:03 crc kubenswrapper[4934]: E1002 10:12:03.387471 4934 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfb4baab_8e79_44ed_8d2a_7da87c069cc2.slice/crio-e8337a262ee7aad19d59cecd8c1c7a0a2b5aa57a2cde4ecdc77477461bfc5d1e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfb4baab_8e79_44ed_8d2a_7da87c069cc2.slice/crio-conmon-e8337a262ee7aad19d59cecd8c1c7a0a2b5aa57a2cde4ecdc77477461bfc5d1e.scope\": RecentStats: unable to find data in memory cache]" Oct 02 10:12:10 crc kubenswrapper[4934]: I1002 10:12:10.499257 4934 scope.go:117] "RemoveContainer" containerID="b74f83fb8d92680f011c564f24d2339dcb490fefe27cf43645816a56cbb27d57" Oct 02 10:12:10 crc kubenswrapper[4934]: I1002 10:12:10.536473 4934 scope.go:117] "RemoveContainer" containerID="7826f59cb7f15b3e25a5cbac82bed398ef57ad483fbbed25d37b37218b482cf8" Oct 02 10:12:10 crc kubenswrapper[4934]: I1002 10:12:10.571369 4934 scope.go:117] "RemoveContainer" containerID="3c4a89f33ba516b7c1f812218bd4d35a373ceb64cc4479e97e705b984e4610db" Oct 02 10:12:10 crc kubenswrapper[4934]: I1002 10:12:10.613524 4934 scope.go:117] "RemoveContainer" containerID="09ddc37d1aa16f33abd00be6bd097bcc1c45216a18e235121272e97806bb67e7" Oct 02 10:12:13 crc kubenswrapper[4934]: E1002 10:12:13.634970 4934 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfb4baab_8e79_44ed_8d2a_7da87c069cc2.slice/crio-conmon-e8337a262ee7aad19d59cecd8c1c7a0a2b5aa57a2cde4ecdc77477461bfc5d1e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfb4baab_8e79_44ed_8d2a_7da87c069cc2.slice/crio-e8337a262ee7aad19d59cecd8c1c7a0a2b5aa57a2cde4ecdc77477461bfc5d1e.scope\": RecentStats: unable to find data in memory cache]" Oct 02 10:12:23 crc kubenswrapper[4934]: E1002 10:12:23.879822 4934 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfb4baab_8e79_44ed_8d2a_7da87c069cc2.slice/crio-conmon-e8337a262ee7aad19d59cecd8c1c7a0a2b5aa57a2cde4ecdc77477461bfc5d1e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfb4baab_8e79_44ed_8d2a_7da87c069cc2.slice/crio-e8337a262ee7aad19d59cecd8c1c7a0a2b5aa57a2cde4ecdc77477461bfc5d1e.scope\": RecentStats: unable to find data in memory cache]" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.311113 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rk955"] Oct 02 10:12:31 crc kubenswrapper[4934]: E1002 10:12:31.312405 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf007c3a-2ba5-4d94-a789-a5413787d6d8" containerName="mariadb-account-delete" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.312434 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf007c3a-2ba5-4d94-a789-a5413787d6d8" containerName="mariadb-account-delete" Oct 02 10:12:31 crc kubenswrapper[4934]: E1002 10:12:31.312468 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="account-replicator" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.312484 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="account-replicator" Oct 02 10:12:31 crc kubenswrapper[4934]: E1002 10:12:31.312506 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfb4baab-8e79-44ed-8d2a-7da87c069cc2" containerName="extract-content" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.312522 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfb4baab-8e79-44ed-8d2a-7da87c069cc2" containerName="extract-content" Oct 02 10:12:31 crc kubenswrapper[4934]: E1002 10:12:31.312551 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1146cda2-78fc-40c5-b0d2-f8c19368fa2c" containerName="ovsdb-server-init" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.312568 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1146cda2-78fc-40c5-b0d2-f8c19368fa2c" containerName="ovsdb-server-init" Oct 02 10:12:31 crc kubenswrapper[4934]: E1002 10:12:31.312635 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfb4baab-8e79-44ed-8d2a-7da87c069cc2" containerName="registry-server" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.312655 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfb4baab-8e79-44ed-8d2a-7da87c069cc2" containerName="registry-server" Oct 02 10:12:31 crc kubenswrapper[4934]: E1002 10:12:31.312687 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="container-auditor" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.312703 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="container-auditor" Oct 02 10:12:31 crc kubenswrapper[4934]: E1002 10:12:31.312730 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="object-updater" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.312745 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="object-updater" Oct 02 10:12:31 crc kubenswrapper[4934]: E1002 10:12:31.312773 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="account-server" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.312788 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="account-server" Oct 02 10:12:31 crc kubenswrapper[4934]: E1002 10:12:31.312818 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7f7a4d6-c246-4c88-85e1-573fd07ec33f" containerName="mariadb-account-delete" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.312832 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7f7a4d6-c246-4c88-85e1-573fd07ec33f" containerName="mariadb-account-delete" Oct 02 10:12:31 crc kubenswrapper[4934]: E1002 10:12:31.312853 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="object-expirer" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.312867 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="object-expirer" Oct 02 10:12:31 crc kubenswrapper[4934]: E1002 10:12:31.312884 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="rsync" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.312899 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="rsync" Oct 02 10:12:31 crc kubenswrapper[4934]: E1002 10:12:31.312921 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="container-server" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.312937 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="container-server" Oct 02 10:12:31 crc kubenswrapper[4934]: E1002 10:12:31.312953 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8679bb63-5c56-42a3-9b2c-42303e3afef6" containerName="mariadb-account-delete" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.312970 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8679bb63-5c56-42a3-9b2c-42303e3afef6" containerName="mariadb-account-delete" Oct 02 10:12:31 crc kubenswrapper[4934]: E1002 10:12:31.312997 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfb4baab-8e79-44ed-8d2a-7da87c069cc2" containerName="extract-utilities" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.313014 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfb4baab-8e79-44ed-8d2a-7da87c069cc2" containerName="extract-utilities" Oct 02 10:12:31 crc kubenswrapper[4934]: E1002 10:12:31.313090 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="object-auditor" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.313108 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="object-auditor" Oct 02 10:12:31 crc kubenswrapper[4934]: E1002 10:12:31.313139 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="swift-recon-cron" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.313155 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="swift-recon-cron" Oct 02 10:12:31 crc kubenswrapper[4934]: E1002 10:12:31.313180 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1146cda2-78fc-40c5-b0d2-f8c19368fa2c" containerName="ovs-vswitchd" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.313196 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1146cda2-78fc-40c5-b0d2-f8c19368fa2c" containerName="ovs-vswitchd" Oct 02 10:12:31 crc kubenswrapper[4934]: E1002 10:12:31.313220 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="container-replicator" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.313235 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="container-replicator" Oct 02 10:12:31 crc kubenswrapper[4934]: E1002 10:12:31.313261 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="object-replicator" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.313276 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="object-replicator" Oct 02 10:12:31 crc kubenswrapper[4934]: E1002 10:12:31.313300 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="container-updater" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.313316 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="container-updater" Oct 02 10:12:31 crc kubenswrapper[4934]: E1002 10:12:31.313342 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="object-server" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.313357 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="object-server" Oct 02 10:12:31 crc kubenswrapper[4934]: E1002 10:12:31.313383 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="account-auditor" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.313400 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="account-auditor" Oct 02 10:12:31 crc kubenswrapper[4934]: E1002 10:12:31.313423 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1146cda2-78fc-40c5-b0d2-f8c19368fa2c" containerName="ovsdb-server" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.313438 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1146cda2-78fc-40c5-b0d2-f8c19368fa2c" containerName="ovsdb-server" Oct 02 10:12:31 crc kubenswrapper[4934]: E1002 10:12:31.313464 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94a95a16-6bdd-4c08-8340-13fd85455b7a" containerName="probe" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.313481 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="94a95a16-6bdd-4c08-8340-13fd85455b7a" containerName="probe" Oct 02 10:12:31 crc kubenswrapper[4934]: E1002 10:12:31.313502 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ceb70f8-5cc3-4e91-9bc1-6fd0b7575efc" containerName="mariadb-account-delete" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.313517 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ceb70f8-5cc3-4e91-9bc1-6fd0b7575efc" containerName="mariadb-account-delete" Oct 02 10:12:31 crc kubenswrapper[4934]: E1002 10:12:31.313545 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94a95a16-6bdd-4c08-8340-13fd85455b7a" containerName="cinder-scheduler" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.313561 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="94a95a16-6bdd-4c08-8340-13fd85455b7a" containerName="cinder-scheduler" Oct 02 10:12:31 crc kubenswrapper[4934]: E1002 10:12:31.313627 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="account-reaper" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.313644 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="account-reaper" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.313956 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ceb70f8-5cc3-4e91-9bc1-6fd0b7575efc" containerName="mariadb-account-delete" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.313992 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="object-updater" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.314010 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="94a95a16-6bdd-4c08-8340-13fd85455b7a" containerName="probe" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.314037 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="container-updater" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.314055 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="swift-recon-cron" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.314071 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="account-replicator" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.314089 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="account-reaper" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.314109 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="account-server" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.314141 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="rsync" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.314159 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="object-replicator" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.314176 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="object-server" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.314201 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="container-replicator" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.314226 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="8679bb63-5c56-42a3-9b2c-42303e3afef6" containerName="mariadb-account-delete" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.314242 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="object-auditor" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.314267 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="account-auditor" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.314296 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1146cda2-78fc-40c5-b0d2-f8c19368fa2c" containerName="ovsdb-server" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.314322 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf007c3a-2ba5-4d94-a789-a5413787d6d8" containerName="mariadb-account-delete" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.314351 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="object-expirer" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.314370 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="container-auditor" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.314387 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7f7a4d6-c246-4c88-85e1-573fd07ec33f" containerName="mariadb-account-delete" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.314408 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1146cda2-78fc-40c5-b0d2-f8c19368fa2c" containerName="ovs-vswitchd" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.314437 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cebd152-5aaf-4543-be11-7ba327cfdf0e" containerName="container-server" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.314465 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfb4baab-8e79-44ed-8d2a-7da87c069cc2" containerName="registry-server" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.314489 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="94a95a16-6bdd-4c08-8340-13fd85455b7a" containerName="cinder-scheduler" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.316968 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rk955" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.330849 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rk955"] Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.445522 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bf54571-c255-4812-9e6f-1de956ca190c-catalog-content\") pod \"redhat-marketplace-rk955\" (UID: \"6bf54571-c255-4812-9e6f-1de956ca190c\") " pod="openshift-marketplace/redhat-marketplace-rk955" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.445754 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bf54571-c255-4812-9e6f-1de956ca190c-utilities\") pod \"redhat-marketplace-rk955\" (UID: \"6bf54571-c255-4812-9e6f-1de956ca190c\") " pod="openshift-marketplace/redhat-marketplace-rk955" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.445824 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgm2h\" (UniqueName: \"kubernetes.io/projected/6bf54571-c255-4812-9e6f-1de956ca190c-kube-api-access-mgm2h\") pod \"redhat-marketplace-rk955\" (UID: \"6bf54571-c255-4812-9e6f-1de956ca190c\") " pod="openshift-marketplace/redhat-marketplace-rk955" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.546992 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bf54571-c255-4812-9e6f-1de956ca190c-catalog-content\") pod \"redhat-marketplace-rk955\" (UID: \"6bf54571-c255-4812-9e6f-1de956ca190c\") " pod="openshift-marketplace/redhat-marketplace-rk955" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.547076 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bf54571-c255-4812-9e6f-1de956ca190c-utilities\") pod \"redhat-marketplace-rk955\" (UID: \"6bf54571-c255-4812-9e6f-1de956ca190c\") " pod="openshift-marketplace/redhat-marketplace-rk955" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.547115 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgm2h\" (UniqueName: \"kubernetes.io/projected/6bf54571-c255-4812-9e6f-1de956ca190c-kube-api-access-mgm2h\") pod \"redhat-marketplace-rk955\" (UID: \"6bf54571-c255-4812-9e6f-1de956ca190c\") " pod="openshift-marketplace/redhat-marketplace-rk955" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.547540 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bf54571-c255-4812-9e6f-1de956ca190c-catalog-content\") pod \"redhat-marketplace-rk955\" (UID: \"6bf54571-c255-4812-9e6f-1de956ca190c\") " pod="openshift-marketplace/redhat-marketplace-rk955" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.547630 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bf54571-c255-4812-9e6f-1de956ca190c-utilities\") pod \"redhat-marketplace-rk955\" (UID: \"6bf54571-c255-4812-9e6f-1de956ca190c\") " pod="openshift-marketplace/redhat-marketplace-rk955" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.570410 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgm2h\" (UniqueName: \"kubernetes.io/projected/6bf54571-c255-4812-9e6f-1de956ca190c-kube-api-access-mgm2h\") pod \"redhat-marketplace-rk955\" (UID: \"6bf54571-c255-4812-9e6f-1de956ca190c\") " pod="openshift-marketplace/redhat-marketplace-rk955" Oct 02 10:12:31 crc kubenswrapper[4934]: I1002 10:12:31.647354 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rk955" Oct 02 10:12:32 crc kubenswrapper[4934]: I1002 10:12:32.101639 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rk955"] Oct 02 10:12:32 crc kubenswrapper[4934]: I1002 10:12:32.366847 4934 generic.go:334] "Generic (PLEG): container finished" podID="6bf54571-c255-4812-9e6f-1de956ca190c" containerID="61e08de558d553fafc610ba68e1e2c56901fffe62b7e5b157cb324caad8ccd9d" exitCode=0 Oct 02 10:12:32 crc kubenswrapper[4934]: I1002 10:12:32.366934 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rk955" event={"ID":"6bf54571-c255-4812-9e6f-1de956ca190c","Type":"ContainerDied","Data":"61e08de558d553fafc610ba68e1e2c56901fffe62b7e5b157cb324caad8ccd9d"} Oct 02 10:12:32 crc kubenswrapper[4934]: I1002 10:12:32.367348 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rk955" event={"ID":"6bf54571-c255-4812-9e6f-1de956ca190c","Type":"ContainerStarted","Data":"6be59dafcfa36bbbf640e2e245829a57d11486016340d84681934640b4a54c38"} Oct 02 10:12:33 crc kubenswrapper[4934]: I1002 10:12:33.377706 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rk955" event={"ID":"6bf54571-c255-4812-9e6f-1de956ca190c","Type":"ContainerStarted","Data":"bc266cbbf1476dcf00e827a955d9402bb1f4f8982fe97985d8211e312dce6901"} Oct 02 10:12:34 crc kubenswrapper[4934]: E1002 10:12:34.103865 4934 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfb4baab_8e79_44ed_8d2a_7da87c069cc2.slice/crio-e8337a262ee7aad19d59cecd8c1c7a0a2b5aa57a2cde4ecdc77477461bfc5d1e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfb4baab_8e79_44ed_8d2a_7da87c069cc2.slice/crio-conmon-e8337a262ee7aad19d59cecd8c1c7a0a2b5aa57a2cde4ecdc77477461bfc5d1e.scope\": RecentStats: unable to find data in memory cache]" Oct 02 10:12:34 crc kubenswrapper[4934]: I1002 10:12:34.386420 4934 generic.go:334] "Generic (PLEG): container finished" podID="6bf54571-c255-4812-9e6f-1de956ca190c" containerID="bc266cbbf1476dcf00e827a955d9402bb1f4f8982fe97985d8211e312dce6901" exitCode=0 Oct 02 10:12:34 crc kubenswrapper[4934]: I1002 10:12:34.386472 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rk955" event={"ID":"6bf54571-c255-4812-9e6f-1de956ca190c","Type":"ContainerDied","Data":"bc266cbbf1476dcf00e827a955d9402bb1f4f8982fe97985d8211e312dce6901"} Oct 02 10:12:35 crc kubenswrapper[4934]: I1002 10:12:35.396349 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rk955" event={"ID":"6bf54571-c255-4812-9e6f-1de956ca190c","Type":"ContainerStarted","Data":"fdf406482c153d2fce3fb69a315464a48a4371719e0f3afbc20e8e298dc61908"} Oct 02 10:12:35 crc kubenswrapper[4934]: I1002 10:12:35.427283 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rk955" podStartSLOduration=2.000414755 podStartE2EDuration="4.4272642s" podCreationTimestamp="2025-10-02 10:12:31 +0000 UTC" firstStartedPulling="2025-10-02 10:12:32.369049564 +0000 UTC m=+1424.121691096" lastFinishedPulling="2025-10-02 10:12:34.795899019 +0000 UTC m=+1426.548540541" observedRunningTime="2025-10-02 10:12:35.420865121 +0000 UTC m=+1427.173506653" watchObservedRunningTime="2025-10-02 10:12:35.4272642 +0000 UTC m=+1427.179905732" Oct 02 10:12:41 crc kubenswrapper[4934]: I1002 10:12:41.648573 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rk955" Oct 02 10:12:41 crc kubenswrapper[4934]: I1002 10:12:41.649180 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rk955" Oct 02 10:12:41 crc kubenswrapper[4934]: I1002 10:12:41.707414 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rk955" Oct 02 10:12:42 crc kubenswrapper[4934]: I1002 10:12:42.549292 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rk955" Oct 02 10:12:42 crc kubenswrapper[4934]: I1002 10:12:42.597786 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rk955"] Oct 02 10:12:44 crc kubenswrapper[4934]: E1002 10:12:44.372989 4934 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfb4baab_8e79_44ed_8d2a_7da87c069cc2.slice/crio-e8337a262ee7aad19d59cecd8c1c7a0a2b5aa57a2cde4ecdc77477461bfc5d1e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfb4baab_8e79_44ed_8d2a_7da87c069cc2.slice/crio-conmon-e8337a262ee7aad19d59cecd8c1c7a0a2b5aa57a2cde4ecdc77477461bfc5d1e.scope\": RecentStats: unable to find data in memory cache]" Oct 02 10:12:44 crc kubenswrapper[4934]: I1002 10:12:44.501145 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rk955" podUID="6bf54571-c255-4812-9e6f-1de956ca190c" containerName="registry-server" containerID="cri-o://fdf406482c153d2fce3fb69a315464a48a4371719e0f3afbc20e8e298dc61908" gracePeriod=2 Oct 02 10:12:45 crc kubenswrapper[4934]: I1002 10:12:45.006296 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rk955" Oct 02 10:12:45 crc kubenswrapper[4934]: I1002 10:12:45.079915 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bf54571-c255-4812-9e6f-1de956ca190c-catalog-content\") pod \"6bf54571-c255-4812-9e6f-1de956ca190c\" (UID: \"6bf54571-c255-4812-9e6f-1de956ca190c\") " Oct 02 10:12:45 crc kubenswrapper[4934]: I1002 10:12:45.080821 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bf54571-c255-4812-9e6f-1de956ca190c-utilities\") pod \"6bf54571-c255-4812-9e6f-1de956ca190c\" (UID: \"6bf54571-c255-4812-9e6f-1de956ca190c\") " Oct 02 10:12:45 crc kubenswrapper[4934]: I1002 10:12:45.080894 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgm2h\" (UniqueName: \"kubernetes.io/projected/6bf54571-c255-4812-9e6f-1de956ca190c-kube-api-access-mgm2h\") pod \"6bf54571-c255-4812-9e6f-1de956ca190c\" (UID: \"6bf54571-c255-4812-9e6f-1de956ca190c\") " Oct 02 10:12:45 crc kubenswrapper[4934]: I1002 10:12:45.081837 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6bf54571-c255-4812-9e6f-1de956ca190c-utilities" (OuterVolumeSpecName: "utilities") pod "6bf54571-c255-4812-9e6f-1de956ca190c" (UID: "6bf54571-c255-4812-9e6f-1de956ca190c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:12:45 crc kubenswrapper[4934]: I1002 10:12:45.082973 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bf54571-c255-4812-9e6f-1de956ca190c-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:12:45 crc kubenswrapper[4934]: I1002 10:12:45.089773 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bf54571-c255-4812-9e6f-1de956ca190c-kube-api-access-mgm2h" (OuterVolumeSpecName: "kube-api-access-mgm2h") pod "6bf54571-c255-4812-9e6f-1de956ca190c" (UID: "6bf54571-c255-4812-9e6f-1de956ca190c"). InnerVolumeSpecName "kube-api-access-mgm2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:12:45 crc kubenswrapper[4934]: I1002 10:12:45.098094 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6bf54571-c255-4812-9e6f-1de956ca190c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6bf54571-c255-4812-9e6f-1de956ca190c" (UID: "6bf54571-c255-4812-9e6f-1de956ca190c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:12:45 crc kubenswrapper[4934]: I1002 10:12:45.185996 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bf54571-c255-4812-9e6f-1de956ca190c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:12:45 crc kubenswrapper[4934]: I1002 10:12:45.187203 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgm2h\" (UniqueName: \"kubernetes.io/projected/6bf54571-c255-4812-9e6f-1de956ca190c-kube-api-access-mgm2h\") on node \"crc\" DevicePath \"\"" Oct 02 10:12:45 crc kubenswrapper[4934]: I1002 10:12:45.517644 4934 generic.go:334] "Generic (PLEG): container finished" podID="6bf54571-c255-4812-9e6f-1de956ca190c" containerID="fdf406482c153d2fce3fb69a315464a48a4371719e0f3afbc20e8e298dc61908" exitCode=0 Oct 02 10:12:45 crc kubenswrapper[4934]: I1002 10:12:45.517701 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rk955" event={"ID":"6bf54571-c255-4812-9e6f-1de956ca190c","Type":"ContainerDied","Data":"fdf406482c153d2fce3fb69a315464a48a4371719e0f3afbc20e8e298dc61908"} Oct 02 10:12:45 crc kubenswrapper[4934]: I1002 10:12:45.517710 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rk955" Oct 02 10:12:45 crc kubenswrapper[4934]: I1002 10:12:45.517742 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rk955" event={"ID":"6bf54571-c255-4812-9e6f-1de956ca190c","Type":"ContainerDied","Data":"6be59dafcfa36bbbf640e2e245829a57d11486016340d84681934640b4a54c38"} Oct 02 10:12:45 crc kubenswrapper[4934]: I1002 10:12:45.517773 4934 scope.go:117] "RemoveContainer" containerID="fdf406482c153d2fce3fb69a315464a48a4371719e0f3afbc20e8e298dc61908" Oct 02 10:12:45 crc kubenswrapper[4934]: I1002 10:12:45.571844 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rk955"] Oct 02 10:12:45 crc kubenswrapper[4934]: I1002 10:12:45.574384 4934 scope.go:117] "RemoveContainer" containerID="bc266cbbf1476dcf00e827a955d9402bb1f4f8982fe97985d8211e312dce6901" Oct 02 10:12:45 crc kubenswrapper[4934]: I1002 10:12:45.581389 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rk955"] Oct 02 10:12:45 crc kubenswrapper[4934]: I1002 10:12:45.601505 4934 scope.go:117] "RemoveContainer" containerID="61e08de558d553fafc610ba68e1e2c56901fffe62b7e5b157cb324caad8ccd9d" Oct 02 10:12:45 crc kubenswrapper[4934]: I1002 10:12:45.644306 4934 scope.go:117] "RemoveContainer" containerID="fdf406482c153d2fce3fb69a315464a48a4371719e0f3afbc20e8e298dc61908" Oct 02 10:12:45 crc kubenswrapper[4934]: E1002 10:12:45.645076 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdf406482c153d2fce3fb69a315464a48a4371719e0f3afbc20e8e298dc61908\": container with ID starting with fdf406482c153d2fce3fb69a315464a48a4371719e0f3afbc20e8e298dc61908 not found: ID does not exist" containerID="fdf406482c153d2fce3fb69a315464a48a4371719e0f3afbc20e8e298dc61908" Oct 02 10:12:45 crc kubenswrapper[4934]: I1002 10:12:45.645134 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdf406482c153d2fce3fb69a315464a48a4371719e0f3afbc20e8e298dc61908"} err="failed to get container status \"fdf406482c153d2fce3fb69a315464a48a4371719e0f3afbc20e8e298dc61908\": rpc error: code = NotFound desc = could not find container \"fdf406482c153d2fce3fb69a315464a48a4371719e0f3afbc20e8e298dc61908\": container with ID starting with fdf406482c153d2fce3fb69a315464a48a4371719e0f3afbc20e8e298dc61908 not found: ID does not exist" Oct 02 10:12:45 crc kubenswrapper[4934]: I1002 10:12:45.645171 4934 scope.go:117] "RemoveContainer" containerID="bc266cbbf1476dcf00e827a955d9402bb1f4f8982fe97985d8211e312dce6901" Oct 02 10:12:45 crc kubenswrapper[4934]: E1002 10:12:45.645628 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc266cbbf1476dcf00e827a955d9402bb1f4f8982fe97985d8211e312dce6901\": container with ID starting with bc266cbbf1476dcf00e827a955d9402bb1f4f8982fe97985d8211e312dce6901 not found: ID does not exist" containerID="bc266cbbf1476dcf00e827a955d9402bb1f4f8982fe97985d8211e312dce6901" Oct 02 10:12:45 crc kubenswrapper[4934]: I1002 10:12:45.645683 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc266cbbf1476dcf00e827a955d9402bb1f4f8982fe97985d8211e312dce6901"} err="failed to get container status \"bc266cbbf1476dcf00e827a955d9402bb1f4f8982fe97985d8211e312dce6901\": rpc error: code = NotFound desc = could not find container \"bc266cbbf1476dcf00e827a955d9402bb1f4f8982fe97985d8211e312dce6901\": container with ID starting with bc266cbbf1476dcf00e827a955d9402bb1f4f8982fe97985d8211e312dce6901 not found: ID does not exist" Oct 02 10:12:45 crc kubenswrapper[4934]: I1002 10:12:45.645715 4934 scope.go:117] "RemoveContainer" containerID="61e08de558d553fafc610ba68e1e2c56901fffe62b7e5b157cb324caad8ccd9d" Oct 02 10:12:45 crc kubenswrapper[4934]: E1002 10:12:45.646075 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61e08de558d553fafc610ba68e1e2c56901fffe62b7e5b157cb324caad8ccd9d\": container with ID starting with 61e08de558d553fafc610ba68e1e2c56901fffe62b7e5b157cb324caad8ccd9d not found: ID does not exist" containerID="61e08de558d553fafc610ba68e1e2c56901fffe62b7e5b157cb324caad8ccd9d" Oct 02 10:12:45 crc kubenswrapper[4934]: I1002 10:12:45.646117 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61e08de558d553fafc610ba68e1e2c56901fffe62b7e5b157cb324caad8ccd9d"} err="failed to get container status \"61e08de558d553fafc610ba68e1e2c56901fffe62b7e5b157cb324caad8ccd9d\": rpc error: code = NotFound desc = could not find container \"61e08de558d553fafc610ba68e1e2c56901fffe62b7e5b157cb324caad8ccd9d\": container with ID starting with 61e08de558d553fafc610ba68e1e2c56901fffe62b7e5b157cb324caad8ccd9d not found: ID does not exist" Oct 02 10:12:46 crc kubenswrapper[4934]: I1002 10:12:46.929834 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bf54571-c255-4812-9e6f-1de956ca190c" path="/var/lib/kubelet/pods/6bf54571-c255-4812-9e6f-1de956ca190c/volumes" Oct 02 10:13:08 crc kubenswrapper[4934]: I1002 10:13:08.439735 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:13:08 crc kubenswrapper[4934]: I1002 10:13:08.440285 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:13:11 crc kubenswrapper[4934]: I1002 10:13:11.434347 4934 scope.go:117] "RemoveContainer" containerID="d6e70f3f0fffee829ba104475e221e24a859707e848e6edd6c64293f59aa17d9" Oct 02 10:13:11 crc kubenswrapper[4934]: I1002 10:13:11.458728 4934 scope.go:117] "RemoveContainer" containerID="fd0b5910cc8779a765c024bbeafee2d307597efe1d65c9d3dbff30dc96cca647" Oct 02 10:13:11 crc kubenswrapper[4934]: I1002 10:13:11.504687 4934 scope.go:117] "RemoveContainer" containerID="2685a9eb90de41e74a709a47f4f8313728ce1ab7fcb3df49f017512f156c070d" Oct 02 10:13:11 crc kubenswrapper[4934]: I1002 10:13:11.533796 4934 scope.go:117] "RemoveContainer" containerID="df3044ab763e8645c1efadc31caa8506372fa484c120cc1970d19fc14e248262" Oct 02 10:13:11 crc kubenswrapper[4934]: I1002 10:13:11.553753 4934 scope.go:117] "RemoveContainer" containerID="2ddbdd41bb6a64a519400ee4292c076412f2dc818cd3e1b0b3fc892184d900bb" Oct 02 10:13:11 crc kubenswrapper[4934]: I1002 10:13:11.583597 4934 scope.go:117] "RemoveContainer" containerID="17fa067adca99d11768b4048a418152c7da822434414f68329addaa440085afb" Oct 02 10:13:11 crc kubenswrapper[4934]: I1002 10:13:11.601478 4934 scope.go:117] "RemoveContainer" containerID="d90d083e790725945807012f42289291311c29e6d6bd45cb9bdc6528aa2de2e4" Oct 02 10:13:11 crc kubenswrapper[4934]: I1002 10:13:11.627997 4934 scope.go:117] "RemoveContainer" containerID="b3acd6151738e0f4de858b51bf54757df9ef9b0a140e37a4f4a15c28e67237be" Oct 02 10:13:11 crc kubenswrapper[4934]: I1002 10:13:11.644881 4934 scope.go:117] "RemoveContainer" containerID="742b3375416bacd7728d2de60efffabe7d330310983b7154b9fa6d01154d25d2" Oct 02 10:13:11 crc kubenswrapper[4934]: I1002 10:13:11.661293 4934 scope.go:117] "RemoveContainer" containerID="8075fe4eb85efceb29e57e7b750c07c1739933c65104eec3beb30a995332f635" Oct 02 10:13:11 crc kubenswrapper[4934]: I1002 10:13:11.690133 4934 scope.go:117] "RemoveContainer" containerID="85cf8b87a7583c3d9bdd7099945fc472bfce2a71819309ed2cd711819382fe15" Oct 02 10:13:11 crc kubenswrapper[4934]: I1002 10:13:11.712790 4934 scope.go:117] "RemoveContainer" containerID="0118a5fe34a7b108765ca45fbcc0401dc71fcf952a3ee604afa4074f90756118" Oct 02 10:13:11 crc kubenswrapper[4934]: I1002 10:13:11.752077 4934 scope.go:117] "RemoveContainer" containerID="5d6e15890d9c5182fa81723619291a5028257eb0f92591a5c0724ba0f008bbbd" Oct 02 10:13:11 crc kubenswrapper[4934]: I1002 10:13:11.780334 4934 scope.go:117] "RemoveContainer" containerID="15c244cad2ff64e103eac7d91d2b751cb8d73809c1159f6386ea5b15b4dae2c9" Oct 02 10:13:11 crc kubenswrapper[4934]: I1002 10:13:11.798008 4934 scope.go:117] "RemoveContainer" containerID="0cda272c203bbc1f03d13a3985bd0215950af56f688bd7b5e079492809f581a7" Oct 02 10:13:38 crc kubenswrapper[4934]: I1002 10:13:38.439762 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:13:38 crc kubenswrapper[4934]: I1002 10:13:38.440927 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:14:08 crc kubenswrapper[4934]: I1002 10:14:08.439487 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:14:08 crc kubenswrapper[4934]: I1002 10:14:08.440155 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:14:08 crc kubenswrapper[4934]: I1002 10:14:08.440214 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 10:14:08 crc kubenswrapper[4934]: I1002 10:14:08.441102 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1"} pod="openshift-machine-config-operator/machine-config-daemon-djh5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 10:14:08 crc kubenswrapper[4934]: I1002 10:14:08.441196 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" containerID="cri-o://f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1" gracePeriod=600 Oct 02 10:14:08 crc kubenswrapper[4934]: E1002 10:14:08.561056 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:14:09 crc kubenswrapper[4934]: I1002 10:14:09.399950 4934 generic.go:334] "Generic (PLEG): container finished" podID="71db06ef-05b0-4f58-b251-b27117a8500a" containerID="f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1" exitCode=0 Oct 02 10:14:09 crc kubenswrapper[4934]: I1002 10:14:09.399997 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerDied","Data":"f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1"} Oct 02 10:14:09 crc kubenswrapper[4934]: I1002 10:14:09.400330 4934 scope.go:117] "RemoveContainer" containerID="577a0f8f7b9147143b952a1ec195b9368082248053d4e85cde27a2757e47ab8c" Oct 02 10:14:09 crc kubenswrapper[4934]: I1002 10:14:09.401151 4934 scope.go:117] "RemoveContainer" containerID="f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1" Oct 02 10:14:09 crc kubenswrapper[4934]: E1002 10:14:09.401519 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:14:12 crc kubenswrapper[4934]: I1002 10:14:12.065398 4934 scope.go:117] "RemoveContainer" containerID="ba2dbd7671783785d215106bb6b057dd861995948702cd2be58811e1a9b4c38e" Oct 02 10:14:12 crc kubenswrapper[4934]: I1002 10:14:12.097174 4934 scope.go:117] "RemoveContainer" containerID="07a3ebc27c86cc0989e83071c71d3c33a754a7f416d67889d187120a9dc11c03" Oct 02 10:14:12 crc kubenswrapper[4934]: I1002 10:14:12.126485 4934 scope.go:117] "RemoveContainer" containerID="1cd0dc5312e4c3b419afaad3821f775ed69083d11a25939daa211e3938162c93" Oct 02 10:14:12 crc kubenswrapper[4934]: I1002 10:14:12.180868 4934 scope.go:117] "RemoveContainer" containerID="65dbab7a03568a76645932c0b29b78b90ef3f3460bb8573a3ba9cd75761a9154" Oct 02 10:14:12 crc kubenswrapper[4934]: I1002 10:14:12.199647 4934 scope.go:117] "RemoveContainer" containerID="3ae22becac0e4812e95a08cb7d2bbcf1e151ec55f7db552c2af6ec857bd58070" Oct 02 10:14:12 crc kubenswrapper[4934]: I1002 10:14:12.218279 4934 scope.go:117] "RemoveContainer" containerID="cee5b6bd1bcc9c2981f3668a02ad37893a0ce6cb09bbaf102043a1606140b34f" Oct 02 10:14:12 crc kubenswrapper[4934]: I1002 10:14:12.238965 4934 scope.go:117] "RemoveContainer" containerID="d4c05b6d8b3d3381710c9dd68250fc8af41ac47622cf6da8cc3795965ae20a68" Oct 02 10:14:12 crc kubenswrapper[4934]: I1002 10:14:12.283404 4934 scope.go:117] "RemoveContainer" containerID="3e597152c8970caf6df4198a19c0fb1ca0108eab7f0f4e07d15ecd0797056ed0" Oct 02 10:14:12 crc kubenswrapper[4934]: I1002 10:14:12.303904 4934 scope.go:117] "RemoveContainer" containerID="c83811ac76f059e3b264882bb3a6859bb99aa9ea06694e42d9967733e316ac15" Oct 02 10:14:12 crc kubenswrapper[4934]: I1002 10:14:12.317495 4934 scope.go:117] "RemoveContainer" containerID="9d35a7ad52afadce99f8b4bb3bb5a4238db85b29282f3f5f2fa42c6bd9b238ef" Oct 02 10:14:12 crc kubenswrapper[4934]: I1002 10:14:12.341909 4934 scope.go:117] "RemoveContainer" containerID="d4c19f695157dd25d239073bba217146b7e1c30d74c166a1697941fd5eab8a53" Oct 02 10:14:12 crc kubenswrapper[4934]: I1002 10:14:12.369855 4934 scope.go:117] "RemoveContainer" containerID="2e7e1ea7602c5dbc0c4396593163b8efff5196b8cba864f827def060e37814e7" Oct 02 10:14:20 crc kubenswrapper[4934]: I1002 10:14:20.913439 4934 scope.go:117] "RemoveContainer" containerID="f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1" Oct 02 10:14:20 crc kubenswrapper[4934]: E1002 10:14:20.914179 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:14:22 crc kubenswrapper[4934]: I1002 10:14:22.214384 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-glj8w"] Oct 02 10:14:22 crc kubenswrapper[4934]: E1002 10:14:22.215318 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bf54571-c255-4812-9e6f-1de956ca190c" containerName="extract-content" Oct 02 10:14:22 crc kubenswrapper[4934]: I1002 10:14:22.215342 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bf54571-c255-4812-9e6f-1de956ca190c" containerName="extract-content" Oct 02 10:14:22 crc kubenswrapper[4934]: E1002 10:14:22.215405 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bf54571-c255-4812-9e6f-1de956ca190c" containerName="registry-server" Oct 02 10:14:22 crc kubenswrapper[4934]: I1002 10:14:22.215419 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bf54571-c255-4812-9e6f-1de956ca190c" containerName="registry-server" Oct 02 10:14:22 crc kubenswrapper[4934]: E1002 10:14:22.215444 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bf54571-c255-4812-9e6f-1de956ca190c" containerName="extract-utilities" Oct 02 10:14:22 crc kubenswrapper[4934]: I1002 10:14:22.215458 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bf54571-c255-4812-9e6f-1de956ca190c" containerName="extract-utilities" Oct 02 10:14:22 crc kubenswrapper[4934]: I1002 10:14:22.215822 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bf54571-c255-4812-9e6f-1de956ca190c" containerName="registry-server" Oct 02 10:14:22 crc kubenswrapper[4934]: I1002 10:14:22.217788 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-glj8w" Oct 02 10:14:22 crc kubenswrapper[4934]: I1002 10:14:22.220405 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-glj8w"] Oct 02 10:14:22 crc kubenswrapper[4934]: I1002 10:14:22.297024 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2627\" (UniqueName: \"kubernetes.io/projected/b3fa4aec-9edd-466a-8388-f6d880c85b6a-kube-api-access-r2627\") pod \"certified-operators-glj8w\" (UID: \"b3fa4aec-9edd-466a-8388-f6d880c85b6a\") " pod="openshift-marketplace/certified-operators-glj8w" Oct 02 10:14:22 crc kubenswrapper[4934]: I1002 10:14:22.297109 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3fa4aec-9edd-466a-8388-f6d880c85b6a-utilities\") pod \"certified-operators-glj8w\" (UID: \"b3fa4aec-9edd-466a-8388-f6d880c85b6a\") " pod="openshift-marketplace/certified-operators-glj8w" Oct 02 10:14:22 crc kubenswrapper[4934]: I1002 10:14:22.297138 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3fa4aec-9edd-466a-8388-f6d880c85b6a-catalog-content\") pod \"certified-operators-glj8w\" (UID: \"b3fa4aec-9edd-466a-8388-f6d880c85b6a\") " pod="openshift-marketplace/certified-operators-glj8w" Oct 02 10:14:22 crc kubenswrapper[4934]: I1002 10:14:22.398641 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3fa4aec-9edd-466a-8388-f6d880c85b6a-utilities\") pod \"certified-operators-glj8w\" (UID: \"b3fa4aec-9edd-466a-8388-f6d880c85b6a\") " pod="openshift-marketplace/certified-operators-glj8w" Oct 02 10:14:22 crc kubenswrapper[4934]: I1002 10:14:22.398710 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3fa4aec-9edd-466a-8388-f6d880c85b6a-catalog-content\") pod \"certified-operators-glj8w\" (UID: \"b3fa4aec-9edd-466a-8388-f6d880c85b6a\") " pod="openshift-marketplace/certified-operators-glj8w" Oct 02 10:14:22 crc kubenswrapper[4934]: I1002 10:14:22.398792 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2627\" (UniqueName: \"kubernetes.io/projected/b3fa4aec-9edd-466a-8388-f6d880c85b6a-kube-api-access-r2627\") pod \"certified-operators-glj8w\" (UID: \"b3fa4aec-9edd-466a-8388-f6d880c85b6a\") " pod="openshift-marketplace/certified-operators-glj8w" Oct 02 10:14:22 crc kubenswrapper[4934]: I1002 10:14:22.399230 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3fa4aec-9edd-466a-8388-f6d880c85b6a-utilities\") pod \"certified-operators-glj8w\" (UID: \"b3fa4aec-9edd-466a-8388-f6d880c85b6a\") " pod="openshift-marketplace/certified-operators-glj8w" Oct 02 10:14:22 crc kubenswrapper[4934]: I1002 10:14:22.399327 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3fa4aec-9edd-466a-8388-f6d880c85b6a-catalog-content\") pod \"certified-operators-glj8w\" (UID: \"b3fa4aec-9edd-466a-8388-f6d880c85b6a\") " pod="openshift-marketplace/certified-operators-glj8w" Oct 02 10:14:22 crc kubenswrapper[4934]: I1002 10:14:22.420577 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2627\" (UniqueName: \"kubernetes.io/projected/b3fa4aec-9edd-466a-8388-f6d880c85b6a-kube-api-access-r2627\") pod \"certified-operators-glj8w\" (UID: \"b3fa4aec-9edd-466a-8388-f6d880c85b6a\") " pod="openshift-marketplace/certified-operators-glj8w" Oct 02 10:14:22 crc kubenswrapper[4934]: I1002 10:14:22.540672 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-glj8w" Oct 02 10:14:22 crc kubenswrapper[4934]: I1002 10:14:22.857334 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-glj8w"] Oct 02 10:14:23 crc kubenswrapper[4934]: I1002 10:14:23.530834 4934 generic.go:334] "Generic (PLEG): container finished" podID="b3fa4aec-9edd-466a-8388-f6d880c85b6a" containerID="ee7f793b537fa5a94c20c7c446bc01c93b23cc60effe838ca194622f7f142d66" exitCode=0 Oct 02 10:14:23 crc kubenswrapper[4934]: I1002 10:14:23.530929 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-glj8w" event={"ID":"b3fa4aec-9edd-466a-8388-f6d880c85b6a","Type":"ContainerDied","Data":"ee7f793b537fa5a94c20c7c446bc01c93b23cc60effe838ca194622f7f142d66"} Oct 02 10:14:23 crc kubenswrapper[4934]: I1002 10:14:23.534143 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-glj8w" event={"ID":"b3fa4aec-9edd-466a-8388-f6d880c85b6a","Type":"ContainerStarted","Data":"f0d46d3a9df11af49f2449052d3cdedc400affe0f74051aa4a96eb4d0b353864"} Oct 02 10:14:24 crc kubenswrapper[4934]: I1002 10:14:24.542420 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-glj8w" event={"ID":"b3fa4aec-9edd-466a-8388-f6d880c85b6a","Type":"ContainerStarted","Data":"1e4fec80231bd72acb49ea81b52c21fb21cfd91376336c99ea53ee61b396ba9c"} Oct 02 10:14:25 crc kubenswrapper[4934]: I1002 10:14:25.557489 4934 generic.go:334] "Generic (PLEG): container finished" podID="b3fa4aec-9edd-466a-8388-f6d880c85b6a" containerID="1e4fec80231bd72acb49ea81b52c21fb21cfd91376336c99ea53ee61b396ba9c" exitCode=0 Oct 02 10:14:25 crc kubenswrapper[4934]: I1002 10:14:25.557882 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-glj8w" event={"ID":"b3fa4aec-9edd-466a-8388-f6d880c85b6a","Type":"ContainerDied","Data":"1e4fec80231bd72acb49ea81b52c21fb21cfd91376336c99ea53ee61b396ba9c"} Oct 02 10:14:26 crc kubenswrapper[4934]: I1002 10:14:26.567286 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-glj8w" event={"ID":"b3fa4aec-9edd-466a-8388-f6d880c85b6a","Type":"ContainerStarted","Data":"195894ab35c68004fd3a05230a1e3390ea8886d9e980f440d8473689fabe990c"} Oct 02 10:14:26 crc kubenswrapper[4934]: I1002 10:14:26.586659 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-glj8w" podStartSLOduration=2.013377468 podStartE2EDuration="4.586641588s" podCreationTimestamp="2025-10-02 10:14:22 +0000 UTC" firstStartedPulling="2025-10-02 10:14:23.533099051 +0000 UTC m=+1535.285740583" lastFinishedPulling="2025-10-02 10:14:26.106363181 +0000 UTC m=+1537.859004703" observedRunningTime="2025-10-02 10:14:26.580993519 +0000 UTC m=+1538.333635061" watchObservedRunningTime="2025-10-02 10:14:26.586641588 +0000 UTC m=+1538.339283110" Oct 02 10:14:32 crc kubenswrapper[4934]: I1002 10:14:32.541066 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-glj8w" Oct 02 10:14:32 crc kubenswrapper[4934]: I1002 10:14:32.542824 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-glj8w" Oct 02 10:14:32 crc kubenswrapper[4934]: I1002 10:14:32.607034 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-glj8w" Oct 02 10:14:32 crc kubenswrapper[4934]: I1002 10:14:32.666799 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-glj8w" Oct 02 10:14:32 crc kubenswrapper[4934]: I1002 10:14:32.848555 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-glj8w"] Oct 02 10:14:32 crc kubenswrapper[4934]: I1002 10:14:32.913832 4934 scope.go:117] "RemoveContainer" containerID="f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1" Oct 02 10:14:32 crc kubenswrapper[4934]: E1002 10:14:32.914132 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:14:34 crc kubenswrapper[4934]: I1002 10:14:34.644195 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-glj8w" podUID="b3fa4aec-9edd-466a-8388-f6d880c85b6a" containerName="registry-server" containerID="cri-o://195894ab35c68004fd3a05230a1e3390ea8886d9e980f440d8473689fabe990c" gracePeriod=2 Oct 02 10:14:36 crc kubenswrapper[4934]: I1002 10:14:36.439376 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-glj8w" Oct 02 10:14:36 crc kubenswrapper[4934]: I1002 10:14:36.507675 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3fa4aec-9edd-466a-8388-f6d880c85b6a-utilities\") pod \"b3fa4aec-9edd-466a-8388-f6d880c85b6a\" (UID: \"b3fa4aec-9edd-466a-8388-f6d880c85b6a\") " Oct 02 10:14:36 crc kubenswrapper[4934]: I1002 10:14:36.507817 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2627\" (UniqueName: \"kubernetes.io/projected/b3fa4aec-9edd-466a-8388-f6d880c85b6a-kube-api-access-r2627\") pod \"b3fa4aec-9edd-466a-8388-f6d880c85b6a\" (UID: \"b3fa4aec-9edd-466a-8388-f6d880c85b6a\") " Oct 02 10:14:36 crc kubenswrapper[4934]: I1002 10:14:36.507886 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3fa4aec-9edd-466a-8388-f6d880c85b6a-catalog-content\") pod \"b3fa4aec-9edd-466a-8388-f6d880c85b6a\" (UID: \"b3fa4aec-9edd-466a-8388-f6d880c85b6a\") " Oct 02 10:14:36 crc kubenswrapper[4934]: I1002 10:14:36.508725 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3fa4aec-9edd-466a-8388-f6d880c85b6a-utilities" (OuterVolumeSpecName: "utilities") pod "b3fa4aec-9edd-466a-8388-f6d880c85b6a" (UID: "b3fa4aec-9edd-466a-8388-f6d880c85b6a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:14:36 crc kubenswrapper[4934]: I1002 10:14:36.518860 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3fa4aec-9edd-466a-8388-f6d880c85b6a-kube-api-access-r2627" (OuterVolumeSpecName: "kube-api-access-r2627") pod "b3fa4aec-9edd-466a-8388-f6d880c85b6a" (UID: "b3fa4aec-9edd-466a-8388-f6d880c85b6a"). InnerVolumeSpecName "kube-api-access-r2627". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:14:36 crc kubenswrapper[4934]: I1002 10:14:36.552171 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3fa4aec-9edd-466a-8388-f6d880c85b6a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b3fa4aec-9edd-466a-8388-f6d880c85b6a" (UID: "b3fa4aec-9edd-466a-8388-f6d880c85b6a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:14:36 crc kubenswrapper[4934]: I1002 10:14:36.609606 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3fa4aec-9edd-466a-8388-f6d880c85b6a-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:14:36 crc kubenswrapper[4934]: I1002 10:14:36.609667 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2627\" (UniqueName: \"kubernetes.io/projected/b3fa4aec-9edd-466a-8388-f6d880c85b6a-kube-api-access-r2627\") on node \"crc\" DevicePath \"\"" Oct 02 10:14:36 crc kubenswrapper[4934]: I1002 10:14:36.609691 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3fa4aec-9edd-466a-8388-f6d880c85b6a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:14:36 crc kubenswrapper[4934]: I1002 10:14:36.662926 4934 generic.go:334] "Generic (PLEG): container finished" podID="b3fa4aec-9edd-466a-8388-f6d880c85b6a" containerID="195894ab35c68004fd3a05230a1e3390ea8886d9e980f440d8473689fabe990c" exitCode=0 Oct 02 10:14:36 crc kubenswrapper[4934]: I1002 10:14:36.662976 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-glj8w" event={"ID":"b3fa4aec-9edd-466a-8388-f6d880c85b6a","Type":"ContainerDied","Data":"195894ab35c68004fd3a05230a1e3390ea8886d9e980f440d8473689fabe990c"} Oct 02 10:14:36 crc kubenswrapper[4934]: I1002 10:14:36.663023 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-glj8w" Oct 02 10:14:36 crc kubenswrapper[4934]: I1002 10:14:36.663038 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-glj8w" event={"ID":"b3fa4aec-9edd-466a-8388-f6d880c85b6a","Type":"ContainerDied","Data":"f0d46d3a9df11af49f2449052d3cdedc400affe0f74051aa4a96eb4d0b353864"} Oct 02 10:14:36 crc kubenswrapper[4934]: I1002 10:14:36.663061 4934 scope.go:117] "RemoveContainer" containerID="195894ab35c68004fd3a05230a1e3390ea8886d9e980f440d8473689fabe990c" Oct 02 10:14:36 crc kubenswrapper[4934]: I1002 10:14:36.699703 4934 scope.go:117] "RemoveContainer" containerID="1e4fec80231bd72acb49ea81b52c21fb21cfd91376336c99ea53ee61b396ba9c" Oct 02 10:14:36 crc kubenswrapper[4934]: I1002 10:14:36.704719 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-glj8w"] Oct 02 10:14:36 crc kubenswrapper[4934]: I1002 10:14:36.721267 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-glj8w"] Oct 02 10:14:36 crc kubenswrapper[4934]: I1002 10:14:36.726489 4934 scope.go:117] "RemoveContainer" containerID="ee7f793b537fa5a94c20c7c446bc01c93b23cc60effe838ca194622f7f142d66" Oct 02 10:14:36 crc kubenswrapper[4934]: I1002 10:14:36.766335 4934 scope.go:117] "RemoveContainer" containerID="195894ab35c68004fd3a05230a1e3390ea8886d9e980f440d8473689fabe990c" Oct 02 10:14:36 crc kubenswrapper[4934]: E1002 10:14:36.767049 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"195894ab35c68004fd3a05230a1e3390ea8886d9e980f440d8473689fabe990c\": container with ID starting with 195894ab35c68004fd3a05230a1e3390ea8886d9e980f440d8473689fabe990c not found: ID does not exist" containerID="195894ab35c68004fd3a05230a1e3390ea8886d9e980f440d8473689fabe990c" Oct 02 10:14:36 crc kubenswrapper[4934]: I1002 10:14:36.767224 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"195894ab35c68004fd3a05230a1e3390ea8886d9e980f440d8473689fabe990c"} err="failed to get container status \"195894ab35c68004fd3a05230a1e3390ea8886d9e980f440d8473689fabe990c\": rpc error: code = NotFound desc = could not find container \"195894ab35c68004fd3a05230a1e3390ea8886d9e980f440d8473689fabe990c\": container with ID starting with 195894ab35c68004fd3a05230a1e3390ea8886d9e980f440d8473689fabe990c not found: ID does not exist" Oct 02 10:14:36 crc kubenswrapper[4934]: I1002 10:14:36.767350 4934 scope.go:117] "RemoveContainer" containerID="1e4fec80231bd72acb49ea81b52c21fb21cfd91376336c99ea53ee61b396ba9c" Oct 02 10:14:36 crc kubenswrapper[4934]: E1002 10:14:36.768112 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e4fec80231bd72acb49ea81b52c21fb21cfd91376336c99ea53ee61b396ba9c\": container with ID starting with 1e4fec80231bd72acb49ea81b52c21fb21cfd91376336c99ea53ee61b396ba9c not found: ID does not exist" containerID="1e4fec80231bd72acb49ea81b52c21fb21cfd91376336c99ea53ee61b396ba9c" Oct 02 10:14:36 crc kubenswrapper[4934]: I1002 10:14:36.768163 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e4fec80231bd72acb49ea81b52c21fb21cfd91376336c99ea53ee61b396ba9c"} err="failed to get container status \"1e4fec80231bd72acb49ea81b52c21fb21cfd91376336c99ea53ee61b396ba9c\": rpc error: code = NotFound desc = could not find container \"1e4fec80231bd72acb49ea81b52c21fb21cfd91376336c99ea53ee61b396ba9c\": container with ID starting with 1e4fec80231bd72acb49ea81b52c21fb21cfd91376336c99ea53ee61b396ba9c not found: ID does not exist" Oct 02 10:14:36 crc kubenswrapper[4934]: I1002 10:14:36.768193 4934 scope.go:117] "RemoveContainer" containerID="ee7f793b537fa5a94c20c7c446bc01c93b23cc60effe838ca194622f7f142d66" Oct 02 10:14:36 crc kubenswrapper[4934]: E1002 10:14:36.768566 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee7f793b537fa5a94c20c7c446bc01c93b23cc60effe838ca194622f7f142d66\": container with ID starting with ee7f793b537fa5a94c20c7c446bc01c93b23cc60effe838ca194622f7f142d66 not found: ID does not exist" containerID="ee7f793b537fa5a94c20c7c446bc01c93b23cc60effe838ca194622f7f142d66" Oct 02 10:14:36 crc kubenswrapper[4934]: I1002 10:14:36.768614 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee7f793b537fa5a94c20c7c446bc01c93b23cc60effe838ca194622f7f142d66"} err="failed to get container status \"ee7f793b537fa5a94c20c7c446bc01c93b23cc60effe838ca194622f7f142d66\": rpc error: code = NotFound desc = could not find container \"ee7f793b537fa5a94c20c7c446bc01c93b23cc60effe838ca194622f7f142d66\": container with ID starting with ee7f793b537fa5a94c20c7c446bc01c93b23cc60effe838ca194622f7f142d66 not found: ID does not exist" Oct 02 10:14:36 crc kubenswrapper[4934]: E1002 10:14:36.769102 4934 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3fa4aec_9edd_466a_8388_f6d880c85b6a.slice/crio-f0d46d3a9df11af49f2449052d3cdedc400affe0f74051aa4a96eb4d0b353864\": RecentStats: unable to find data in memory cache]" Oct 02 10:14:36 crc kubenswrapper[4934]: I1002 10:14:36.926800 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3fa4aec-9edd-466a-8388-f6d880c85b6a" path="/var/lib/kubelet/pods/b3fa4aec-9edd-466a-8388-f6d880c85b6a/volumes" Oct 02 10:14:45 crc kubenswrapper[4934]: I1002 10:14:45.913533 4934 scope.go:117] "RemoveContainer" containerID="f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1" Oct 02 10:14:45 crc kubenswrapper[4934]: E1002 10:14:45.914318 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:14:56 crc kubenswrapper[4934]: I1002 10:14:56.915033 4934 scope.go:117] "RemoveContainer" containerID="f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1" Oct 02 10:14:56 crc kubenswrapper[4934]: E1002 10:14:56.916095 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:15:00 crc kubenswrapper[4934]: I1002 10:15:00.170124 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323335-9gjf2"] Oct 02 10:15:00 crc kubenswrapper[4934]: E1002 10:15:00.171154 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3fa4aec-9edd-466a-8388-f6d880c85b6a" containerName="extract-utilities" Oct 02 10:15:00 crc kubenswrapper[4934]: I1002 10:15:00.171184 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3fa4aec-9edd-466a-8388-f6d880c85b6a" containerName="extract-utilities" Oct 02 10:15:00 crc kubenswrapper[4934]: E1002 10:15:00.171209 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3fa4aec-9edd-466a-8388-f6d880c85b6a" containerName="registry-server" Oct 02 10:15:00 crc kubenswrapper[4934]: I1002 10:15:00.171220 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3fa4aec-9edd-466a-8388-f6d880c85b6a" containerName="registry-server" Oct 02 10:15:00 crc kubenswrapper[4934]: E1002 10:15:00.171245 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3fa4aec-9edd-466a-8388-f6d880c85b6a" containerName="extract-content" Oct 02 10:15:00 crc kubenswrapper[4934]: I1002 10:15:00.171257 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3fa4aec-9edd-466a-8388-f6d880c85b6a" containerName="extract-content" Oct 02 10:15:00 crc kubenswrapper[4934]: I1002 10:15:00.171489 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3fa4aec-9edd-466a-8388-f6d880c85b6a" containerName="registry-server" Oct 02 10:15:00 crc kubenswrapper[4934]: I1002 10:15:00.172238 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-9gjf2" Oct 02 10:15:00 crc kubenswrapper[4934]: I1002 10:15:00.175109 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 10:15:00 crc kubenswrapper[4934]: I1002 10:15:00.176355 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 10:15:00 crc kubenswrapper[4934]: I1002 10:15:00.180433 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6f4mf\" (UniqueName: \"kubernetes.io/projected/4774f929-979f-4e63-86d2-f01df01c7334-kube-api-access-6f4mf\") pod \"collect-profiles-29323335-9gjf2\" (UID: \"4774f929-979f-4e63-86d2-f01df01c7334\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-9gjf2" Oct 02 10:15:00 crc kubenswrapper[4934]: I1002 10:15:00.180503 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323335-9gjf2"] Oct 02 10:15:00 crc kubenswrapper[4934]: I1002 10:15:00.180513 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4774f929-979f-4e63-86d2-f01df01c7334-config-volume\") pod \"collect-profiles-29323335-9gjf2\" (UID: \"4774f929-979f-4e63-86d2-f01df01c7334\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-9gjf2" Oct 02 10:15:00 crc kubenswrapper[4934]: I1002 10:15:00.180751 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4774f929-979f-4e63-86d2-f01df01c7334-secret-volume\") pod \"collect-profiles-29323335-9gjf2\" (UID: \"4774f929-979f-4e63-86d2-f01df01c7334\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-9gjf2" Oct 02 10:15:00 crc kubenswrapper[4934]: I1002 10:15:00.281838 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6f4mf\" (UniqueName: \"kubernetes.io/projected/4774f929-979f-4e63-86d2-f01df01c7334-kube-api-access-6f4mf\") pod \"collect-profiles-29323335-9gjf2\" (UID: \"4774f929-979f-4e63-86d2-f01df01c7334\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-9gjf2" Oct 02 10:15:00 crc kubenswrapper[4934]: I1002 10:15:00.282136 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4774f929-979f-4e63-86d2-f01df01c7334-config-volume\") pod \"collect-profiles-29323335-9gjf2\" (UID: \"4774f929-979f-4e63-86d2-f01df01c7334\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-9gjf2" Oct 02 10:15:00 crc kubenswrapper[4934]: I1002 10:15:00.282295 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4774f929-979f-4e63-86d2-f01df01c7334-secret-volume\") pod \"collect-profiles-29323335-9gjf2\" (UID: \"4774f929-979f-4e63-86d2-f01df01c7334\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-9gjf2" Oct 02 10:15:00 crc kubenswrapper[4934]: I1002 10:15:00.283439 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4774f929-979f-4e63-86d2-f01df01c7334-config-volume\") pod \"collect-profiles-29323335-9gjf2\" (UID: \"4774f929-979f-4e63-86d2-f01df01c7334\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-9gjf2" Oct 02 10:15:00 crc kubenswrapper[4934]: I1002 10:15:00.292304 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4774f929-979f-4e63-86d2-f01df01c7334-secret-volume\") pod \"collect-profiles-29323335-9gjf2\" (UID: \"4774f929-979f-4e63-86d2-f01df01c7334\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-9gjf2" Oct 02 10:15:00 crc kubenswrapper[4934]: I1002 10:15:00.305149 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6f4mf\" (UniqueName: \"kubernetes.io/projected/4774f929-979f-4e63-86d2-f01df01c7334-kube-api-access-6f4mf\") pod \"collect-profiles-29323335-9gjf2\" (UID: \"4774f929-979f-4e63-86d2-f01df01c7334\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-9gjf2" Oct 02 10:15:00 crc kubenswrapper[4934]: I1002 10:15:00.513461 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-9gjf2" Oct 02 10:15:00 crc kubenswrapper[4934]: I1002 10:15:00.980080 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323335-9gjf2"] Oct 02 10:15:01 crc kubenswrapper[4934]: I1002 10:15:01.916862 4934 generic.go:334] "Generic (PLEG): container finished" podID="4774f929-979f-4e63-86d2-f01df01c7334" containerID="a3b55e537f2b753ac52d41ae11574d8d156a736700427ae68c33484e9521f50b" exitCode=0 Oct 02 10:15:01 crc kubenswrapper[4934]: I1002 10:15:01.916938 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-9gjf2" event={"ID":"4774f929-979f-4e63-86d2-f01df01c7334","Type":"ContainerDied","Data":"a3b55e537f2b753ac52d41ae11574d8d156a736700427ae68c33484e9521f50b"} Oct 02 10:15:01 crc kubenswrapper[4934]: I1002 10:15:01.918085 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-9gjf2" event={"ID":"4774f929-979f-4e63-86d2-f01df01c7334","Type":"ContainerStarted","Data":"b7a16531d0d89d58375639bc2b03aa2bafb564b4d2bad2923c00758d02d18c11"} Oct 02 10:15:03 crc kubenswrapper[4934]: I1002 10:15:03.286036 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-9gjf2" Oct 02 10:15:03 crc kubenswrapper[4934]: I1002 10:15:03.428217 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6f4mf\" (UniqueName: \"kubernetes.io/projected/4774f929-979f-4e63-86d2-f01df01c7334-kube-api-access-6f4mf\") pod \"4774f929-979f-4e63-86d2-f01df01c7334\" (UID: \"4774f929-979f-4e63-86d2-f01df01c7334\") " Oct 02 10:15:03 crc kubenswrapper[4934]: I1002 10:15:03.428388 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4774f929-979f-4e63-86d2-f01df01c7334-config-volume\") pod \"4774f929-979f-4e63-86d2-f01df01c7334\" (UID: \"4774f929-979f-4e63-86d2-f01df01c7334\") " Oct 02 10:15:03 crc kubenswrapper[4934]: I1002 10:15:03.428483 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4774f929-979f-4e63-86d2-f01df01c7334-secret-volume\") pod \"4774f929-979f-4e63-86d2-f01df01c7334\" (UID: \"4774f929-979f-4e63-86d2-f01df01c7334\") " Oct 02 10:15:03 crc kubenswrapper[4934]: I1002 10:15:03.429058 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4774f929-979f-4e63-86d2-f01df01c7334-config-volume" (OuterVolumeSpecName: "config-volume") pod "4774f929-979f-4e63-86d2-f01df01c7334" (UID: "4774f929-979f-4e63-86d2-f01df01c7334"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:15:03 crc kubenswrapper[4934]: I1002 10:15:03.438765 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4774f929-979f-4e63-86d2-f01df01c7334-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4774f929-979f-4e63-86d2-f01df01c7334" (UID: "4774f929-979f-4e63-86d2-f01df01c7334"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:15:03 crc kubenswrapper[4934]: I1002 10:15:03.438831 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4774f929-979f-4e63-86d2-f01df01c7334-kube-api-access-6f4mf" (OuterVolumeSpecName: "kube-api-access-6f4mf") pod "4774f929-979f-4e63-86d2-f01df01c7334" (UID: "4774f929-979f-4e63-86d2-f01df01c7334"). InnerVolumeSpecName "kube-api-access-6f4mf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:15:03 crc kubenswrapper[4934]: I1002 10:15:03.530897 4934 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4774f929-979f-4e63-86d2-f01df01c7334-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 10:15:03 crc kubenswrapper[4934]: I1002 10:15:03.530948 4934 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4774f929-979f-4e63-86d2-f01df01c7334-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 10:15:03 crc kubenswrapper[4934]: I1002 10:15:03.530968 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6f4mf\" (UniqueName: \"kubernetes.io/projected/4774f929-979f-4e63-86d2-f01df01c7334-kube-api-access-6f4mf\") on node \"crc\" DevicePath \"\"" Oct 02 10:15:03 crc kubenswrapper[4934]: I1002 10:15:03.935813 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-9gjf2" event={"ID":"4774f929-979f-4e63-86d2-f01df01c7334","Type":"ContainerDied","Data":"b7a16531d0d89d58375639bc2b03aa2bafb564b4d2bad2923c00758d02d18c11"} Oct 02 10:15:03 crc kubenswrapper[4934]: I1002 10:15:03.935864 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7a16531d0d89d58375639bc2b03aa2bafb564b4d2bad2923c00758d02d18c11" Oct 02 10:15:03 crc kubenswrapper[4934]: I1002 10:15:03.935891 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-9gjf2" Oct 02 10:15:08 crc kubenswrapper[4934]: I1002 10:15:08.921070 4934 scope.go:117] "RemoveContainer" containerID="f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1" Oct 02 10:15:08 crc kubenswrapper[4934]: E1002 10:15:08.922031 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:15:12 crc kubenswrapper[4934]: I1002 10:15:12.478714 4934 scope.go:117] "RemoveContainer" containerID="4b29478d61836822ab583964f5cda288f9e4166121ace0a26a374d08d5e1860a" Oct 02 10:15:12 crc kubenswrapper[4934]: I1002 10:15:12.499435 4934 scope.go:117] "RemoveContainer" containerID="dc9d145452a9c46ec72acd9d97981007f869e0a17f76c8a91b868d88950976eb" Oct 02 10:15:12 crc kubenswrapper[4934]: I1002 10:15:12.541500 4934 scope.go:117] "RemoveContainer" containerID="da7e910ecc8c66f11827ac4b9c15d0fce9bc3238df05f00389ab9585acf7d276" Oct 02 10:15:12 crc kubenswrapper[4934]: I1002 10:15:12.572072 4934 scope.go:117] "RemoveContainer" containerID="e026663f32d10778f738428a7a66f6498bce90ad6030d9b02bb84a6b4e5df0f0" Oct 02 10:15:12 crc kubenswrapper[4934]: I1002 10:15:12.603943 4934 scope.go:117] "RemoveContainer" containerID="be7e01c0579a42586b0d2b4e10c3e188d7aa80a13a9f3bf0750dcee6fa06cd0e" Oct 02 10:15:12 crc kubenswrapper[4934]: I1002 10:15:12.629423 4934 scope.go:117] "RemoveContainer" containerID="2f94b6c5f263f4184b746c3ade707cb59c40c51fe029d679e0ac873fb6c628e2" Oct 02 10:15:12 crc kubenswrapper[4934]: I1002 10:15:12.654455 4934 scope.go:117] "RemoveContainer" containerID="3b14288c234464c443d841b2865a70bd53e55ad242d835bc723d8515de567383" Oct 02 10:15:21 crc kubenswrapper[4934]: I1002 10:15:21.915339 4934 scope.go:117] "RemoveContainer" containerID="f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1" Oct 02 10:15:21 crc kubenswrapper[4934]: E1002 10:15:21.916751 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:15:36 crc kubenswrapper[4934]: I1002 10:15:36.913853 4934 scope.go:117] "RemoveContainer" containerID="f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1" Oct 02 10:15:36 crc kubenswrapper[4934]: E1002 10:15:36.915378 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:15:48 crc kubenswrapper[4934]: I1002 10:15:48.919038 4934 scope.go:117] "RemoveContainer" containerID="f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1" Oct 02 10:15:48 crc kubenswrapper[4934]: E1002 10:15:48.919860 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:16:00 crc kubenswrapper[4934]: I1002 10:16:00.913697 4934 scope.go:117] "RemoveContainer" containerID="f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1" Oct 02 10:16:00 crc kubenswrapper[4934]: E1002 10:16:00.914737 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:16:11 crc kubenswrapper[4934]: I1002 10:16:11.913207 4934 scope.go:117] "RemoveContainer" containerID="f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1" Oct 02 10:16:11 crc kubenswrapper[4934]: E1002 10:16:11.914194 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:16:12 crc kubenswrapper[4934]: I1002 10:16:12.781465 4934 scope.go:117] "RemoveContainer" containerID="6b77fc0e555402c151a111643d602de419a5a7a8fe3ab9b1c50acb80d1af507a" Oct 02 10:16:12 crc kubenswrapper[4934]: I1002 10:16:12.842052 4934 scope.go:117] "RemoveContainer" containerID="b9baba9779ef141c5b9d9cb5a3a64ef9893f0e264bbcdb5320d4acc16116a004" Oct 02 10:16:12 crc kubenswrapper[4934]: I1002 10:16:12.864867 4934 scope.go:117] "RemoveContainer" containerID="8d3c9f48dcfcd16e834077a1b8cefb91be17eaec6d3b75578d4108c226272f38" Oct 02 10:16:24 crc kubenswrapper[4934]: I1002 10:16:24.914079 4934 scope.go:117] "RemoveContainer" containerID="f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1" Oct 02 10:16:24 crc kubenswrapper[4934]: E1002 10:16:24.915224 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:16:38 crc kubenswrapper[4934]: I1002 10:16:38.920623 4934 scope.go:117] "RemoveContainer" containerID="f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1" Oct 02 10:16:38 crc kubenswrapper[4934]: E1002 10:16:38.921736 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:16:53 crc kubenswrapper[4934]: I1002 10:16:53.913218 4934 scope.go:117] "RemoveContainer" containerID="f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1" Oct 02 10:16:53 crc kubenswrapper[4934]: E1002 10:16:53.914219 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:17:08 crc kubenswrapper[4934]: I1002 10:17:08.917680 4934 scope.go:117] "RemoveContainer" containerID="f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1" Oct 02 10:17:08 crc kubenswrapper[4934]: E1002 10:17:08.919383 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:17:12 crc kubenswrapper[4934]: I1002 10:17:12.967495 4934 scope.go:117] "RemoveContainer" containerID="7bf7515a6d0941fe4d48699791654846c0aa70856a9eb0c70058b67cd59fe0cf" Oct 02 10:17:20 crc kubenswrapper[4934]: I1002 10:17:20.913263 4934 scope.go:117] "RemoveContainer" containerID="f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1" Oct 02 10:17:20 crc kubenswrapper[4934]: E1002 10:17:20.914078 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:17:32 crc kubenswrapper[4934]: I1002 10:17:32.914143 4934 scope.go:117] "RemoveContainer" containerID="f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1" Oct 02 10:17:32 crc kubenswrapper[4934]: E1002 10:17:32.916242 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:17:44 crc kubenswrapper[4934]: I1002 10:17:44.913535 4934 scope.go:117] "RemoveContainer" containerID="f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1" Oct 02 10:17:44 crc kubenswrapper[4934]: E1002 10:17:44.914501 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:17:59 crc kubenswrapper[4934]: I1002 10:17:59.913280 4934 scope.go:117] "RemoveContainer" containerID="f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1" Oct 02 10:17:59 crc kubenswrapper[4934]: E1002 10:17:59.914160 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:18:11 crc kubenswrapper[4934]: I1002 10:18:11.912967 4934 scope.go:117] "RemoveContainer" containerID="f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1" Oct 02 10:18:11 crc kubenswrapper[4934]: E1002 10:18:11.914768 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:18:24 crc kubenswrapper[4934]: I1002 10:18:24.914116 4934 scope.go:117] "RemoveContainer" containerID="f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1" Oct 02 10:18:24 crc kubenswrapper[4934]: E1002 10:18:24.915418 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:18:38 crc kubenswrapper[4934]: I1002 10:18:38.921718 4934 scope.go:117] "RemoveContainer" containerID="f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1" Oct 02 10:18:38 crc kubenswrapper[4934]: E1002 10:18:38.922907 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:18:53 crc kubenswrapper[4934]: I1002 10:18:53.913340 4934 scope.go:117] "RemoveContainer" containerID="f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1" Oct 02 10:18:53 crc kubenswrapper[4934]: E1002 10:18:53.914367 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:19:05 crc kubenswrapper[4934]: I1002 10:19:05.912951 4934 scope.go:117] "RemoveContainer" containerID="f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1" Oct 02 10:19:05 crc kubenswrapper[4934]: E1002 10:19:05.913679 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:19:17 crc kubenswrapper[4934]: I1002 10:19:17.912958 4934 scope.go:117] "RemoveContainer" containerID="f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1" Oct 02 10:19:19 crc kubenswrapper[4934]: I1002 10:19:19.214089 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerStarted","Data":"ff97f60f399f8a24942b018acf363427f36922417bcfd215f609a25976781529"} Oct 02 10:21:36 crc kubenswrapper[4934]: I1002 10:21:36.253361 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ttlh2"] Oct 02 10:21:36 crc kubenswrapper[4934]: E1002 10:21:36.254446 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4774f929-979f-4e63-86d2-f01df01c7334" containerName="collect-profiles" Oct 02 10:21:36 crc kubenswrapper[4934]: I1002 10:21:36.254470 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="4774f929-979f-4e63-86d2-f01df01c7334" containerName="collect-profiles" Oct 02 10:21:36 crc kubenswrapper[4934]: I1002 10:21:36.254776 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="4774f929-979f-4e63-86d2-f01df01c7334" containerName="collect-profiles" Oct 02 10:21:36 crc kubenswrapper[4934]: I1002 10:21:36.256503 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ttlh2" Oct 02 10:21:36 crc kubenswrapper[4934]: I1002 10:21:36.284520 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ttlh2"] Oct 02 10:21:36 crc kubenswrapper[4934]: I1002 10:21:36.372776 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38-utilities\") pod \"community-operators-ttlh2\" (UID: \"b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38\") " pod="openshift-marketplace/community-operators-ttlh2" Oct 02 10:21:36 crc kubenswrapper[4934]: I1002 10:21:36.372861 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38-catalog-content\") pod \"community-operators-ttlh2\" (UID: \"b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38\") " pod="openshift-marketplace/community-operators-ttlh2" Oct 02 10:21:36 crc kubenswrapper[4934]: I1002 10:21:36.372925 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swbb9\" (UniqueName: \"kubernetes.io/projected/b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38-kube-api-access-swbb9\") pod \"community-operators-ttlh2\" (UID: \"b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38\") " pod="openshift-marketplace/community-operators-ttlh2" Oct 02 10:21:36 crc kubenswrapper[4934]: I1002 10:21:36.473549 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38-utilities\") pod \"community-operators-ttlh2\" (UID: \"b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38\") " pod="openshift-marketplace/community-operators-ttlh2" Oct 02 10:21:36 crc kubenswrapper[4934]: I1002 10:21:36.473670 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38-catalog-content\") pod \"community-operators-ttlh2\" (UID: \"b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38\") " pod="openshift-marketplace/community-operators-ttlh2" Oct 02 10:21:36 crc kubenswrapper[4934]: I1002 10:21:36.473723 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swbb9\" (UniqueName: \"kubernetes.io/projected/b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38-kube-api-access-swbb9\") pod \"community-operators-ttlh2\" (UID: \"b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38\") " pod="openshift-marketplace/community-operators-ttlh2" Oct 02 10:21:36 crc kubenswrapper[4934]: I1002 10:21:36.474541 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38-utilities\") pod \"community-operators-ttlh2\" (UID: \"b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38\") " pod="openshift-marketplace/community-operators-ttlh2" Oct 02 10:21:36 crc kubenswrapper[4934]: I1002 10:21:36.474919 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38-catalog-content\") pod \"community-operators-ttlh2\" (UID: \"b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38\") " pod="openshift-marketplace/community-operators-ttlh2" Oct 02 10:21:36 crc kubenswrapper[4934]: I1002 10:21:36.497975 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swbb9\" (UniqueName: \"kubernetes.io/projected/b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38-kube-api-access-swbb9\") pod \"community-operators-ttlh2\" (UID: \"b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38\") " pod="openshift-marketplace/community-operators-ttlh2" Oct 02 10:21:36 crc kubenswrapper[4934]: I1002 10:21:36.597614 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ttlh2" Oct 02 10:21:36 crc kubenswrapper[4934]: I1002 10:21:36.911087 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ttlh2"] Oct 02 10:21:37 crc kubenswrapper[4934]: I1002 10:21:37.485989 4934 generic.go:334] "Generic (PLEG): container finished" podID="b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38" containerID="4654c94587cd3a829f74702e7d74bdf3e8598f39c12033775744aed77ecc155b" exitCode=0 Oct 02 10:21:37 crc kubenswrapper[4934]: I1002 10:21:37.486168 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ttlh2" event={"ID":"b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38","Type":"ContainerDied","Data":"4654c94587cd3a829f74702e7d74bdf3e8598f39c12033775744aed77ecc155b"} Oct 02 10:21:37 crc kubenswrapper[4934]: I1002 10:21:37.486350 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ttlh2" event={"ID":"b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38","Type":"ContainerStarted","Data":"b6c638d9bd4bc23f49304b6b1b154f91b0e06fcf50e0b0d5572448c7492314de"} Oct 02 10:21:37 crc kubenswrapper[4934]: I1002 10:21:37.489924 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 10:21:38 crc kubenswrapper[4934]: I1002 10:21:38.439716 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:21:38 crc kubenswrapper[4934]: I1002 10:21:38.440119 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:21:38 crc kubenswrapper[4934]: I1002 10:21:38.500075 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ttlh2" event={"ID":"b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38","Type":"ContainerStarted","Data":"bbb588aba6c3c92d7afe940b0ad573ba0470987c0703f25854c72a727bebe85a"} Oct 02 10:21:39 crc kubenswrapper[4934]: I1002 10:21:39.513959 4934 generic.go:334] "Generic (PLEG): container finished" podID="b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38" containerID="bbb588aba6c3c92d7afe940b0ad573ba0470987c0703f25854c72a727bebe85a" exitCode=0 Oct 02 10:21:39 crc kubenswrapper[4934]: I1002 10:21:39.514026 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ttlh2" event={"ID":"b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38","Type":"ContainerDied","Data":"bbb588aba6c3c92d7afe940b0ad573ba0470987c0703f25854c72a727bebe85a"} Oct 02 10:21:40 crc kubenswrapper[4934]: I1002 10:21:40.528366 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ttlh2" event={"ID":"b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38","Type":"ContainerStarted","Data":"d04d720b31370e962c0c7760ca0cfca1ae3a57a3b98f13b9da8bd54c5b35e81e"} Oct 02 10:21:40 crc kubenswrapper[4934]: I1002 10:21:40.561228 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ttlh2" podStartSLOduration=2.067726297 podStartE2EDuration="4.56120133s" podCreationTimestamp="2025-10-02 10:21:36 +0000 UTC" firstStartedPulling="2025-10-02 10:21:37.488962159 +0000 UTC m=+1969.241603711" lastFinishedPulling="2025-10-02 10:21:39.982437182 +0000 UTC m=+1971.735078744" observedRunningTime="2025-10-02 10:21:40.555424699 +0000 UTC m=+1972.308066261" watchObservedRunningTime="2025-10-02 10:21:40.56120133 +0000 UTC m=+1972.313842882" Oct 02 10:21:46 crc kubenswrapper[4934]: I1002 10:21:46.598601 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ttlh2" Oct 02 10:21:46 crc kubenswrapper[4934]: I1002 10:21:46.599199 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ttlh2" Oct 02 10:21:46 crc kubenswrapper[4934]: I1002 10:21:46.673947 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ttlh2" Oct 02 10:21:47 crc kubenswrapper[4934]: I1002 10:21:47.661321 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ttlh2" Oct 02 10:21:47 crc kubenswrapper[4934]: I1002 10:21:47.756077 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ttlh2"] Oct 02 10:21:49 crc kubenswrapper[4934]: I1002 10:21:49.604806 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ttlh2" podUID="b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38" containerName="registry-server" containerID="cri-o://d04d720b31370e962c0c7760ca0cfca1ae3a57a3b98f13b9da8bd54c5b35e81e" gracePeriod=2 Oct 02 10:21:50 crc kubenswrapper[4934]: I1002 10:21:50.092257 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ttlh2" Oct 02 10:21:50 crc kubenswrapper[4934]: I1002 10:21:50.211382 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38-catalog-content\") pod \"b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38\" (UID: \"b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38\") " Oct 02 10:21:50 crc kubenswrapper[4934]: I1002 10:21:50.211557 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swbb9\" (UniqueName: \"kubernetes.io/projected/b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38-kube-api-access-swbb9\") pod \"b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38\" (UID: \"b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38\") " Oct 02 10:21:50 crc kubenswrapper[4934]: I1002 10:21:50.211651 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38-utilities\") pod \"b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38\" (UID: \"b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38\") " Oct 02 10:21:50 crc kubenswrapper[4934]: I1002 10:21:50.213107 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38-utilities" (OuterVolumeSpecName: "utilities") pod "b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38" (UID: "b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:21:50 crc kubenswrapper[4934]: I1002 10:21:50.220487 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38-kube-api-access-swbb9" (OuterVolumeSpecName: "kube-api-access-swbb9") pod "b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38" (UID: "b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38"). InnerVolumeSpecName "kube-api-access-swbb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:21:50 crc kubenswrapper[4934]: I1002 10:21:50.291192 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38" (UID: "b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:21:50 crc kubenswrapper[4934]: I1002 10:21:50.313192 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:21:50 crc kubenswrapper[4934]: I1002 10:21:50.313239 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swbb9\" (UniqueName: \"kubernetes.io/projected/b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38-kube-api-access-swbb9\") on node \"crc\" DevicePath \"\"" Oct 02 10:21:50 crc kubenswrapper[4934]: I1002 10:21:50.313261 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:21:50 crc kubenswrapper[4934]: I1002 10:21:50.619165 4934 generic.go:334] "Generic (PLEG): container finished" podID="b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38" containerID="d04d720b31370e962c0c7760ca0cfca1ae3a57a3b98f13b9da8bd54c5b35e81e" exitCode=0 Oct 02 10:21:50 crc kubenswrapper[4934]: I1002 10:21:50.619211 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ttlh2" event={"ID":"b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38","Type":"ContainerDied","Data":"d04d720b31370e962c0c7760ca0cfca1ae3a57a3b98f13b9da8bd54c5b35e81e"} Oct 02 10:21:50 crc kubenswrapper[4934]: I1002 10:21:50.619242 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ttlh2" event={"ID":"b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38","Type":"ContainerDied","Data":"b6c638d9bd4bc23f49304b6b1b154f91b0e06fcf50e0b0d5572448c7492314de"} Oct 02 10:21:50 crc kubenswrapper[4934]: I1002 10:21:50.619262 4934 scope.go:117] "RemoveContainer" containerID="d04d720b31370e962c0c7760ca0cfca1ae3a57a3b98f13b9da8bd54c5b35e81e" Oct 02 10:21:50 crc kubenswrapper[4934]: I1002 10:21:50.619334 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ttlh2" Oct 02 10:21:50 crc kubenswrapper[4934]: I1002 10:21:50.651637 4934 scope.go:117] "RemoveContainer" containerID="bbb588aba6c3c92d7afe940b0ad573ba0470987c0703f25854c72a727bebe85a" Oct 02 10:21:50 crc kubenswrapper[4934]: I1002 10:21:50.675290 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ttlh2"] Oct 02 10:21:50 crc kubenswrapper[4934]: I1002 10:21:50.684079 4934 scope.go:117] "RemoveContainer" containerID="4654c94587cd3a829f74702e7d74bdf3e8598f39c12033775744aed77ecc155b" Oct 02 10:21:50 crc kubenswrapper[4934]: I1002 10:21:50.687352 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ttlh2"] Oct 02 10:21:50 crc kubenswrapper[4934]: I1002 10:21:50.714113 4934 scope.go:117] "RemoveContainer" containerID="d04d720b31370e962c0c7760ca0cfca1ae3a57a3b98f13b9da8bd54c5b35e81e" Oct 02 10:21:50 crc kubenswrapper[4934]: E1002 10:21:50.714739 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d04d720b31370e962c0c7760ca0cfca1ae3a57a3b98f13b9da8bd54c5b35e81e\": container with ID starting with d04d720b31370e962c0c7760ca0cfca1ae3a57a3b98f13b9da8bd54c5b35e81e not found: ID does not exist" containerID="d04d720b31370e962c0c7760ca0cfca1ae3a57a3b98f13b9da8bd54c5b35e81e" Oct 02 10:21:50 crc kubenswrapper[4934]: I1002 10:21:50.714810 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d04d720b31370e962c0c7760ca0cfca1ae3a57a3b98f13b9da8bd54c5b35e81e"} err="failed to get container status \"d04d720b31370e962c0c7760ca0cfca1ae3a57a3b98f13b9da8bd54c5b35e81e\": rpc error: code = NotFound desc = could not find container \"d04d720b31370e962c0c7760ca0cfca1ae3a57a3b98f13b9da8bd54c5b35e81e\": container with ID starting with d04d720b31370e962c0c7760ca0cfca1ae3a57a3b98f13b9da8bd54c5b35e81e not found: ID does not exist" Oct 02 10:21:50 crc kubenswrapper[4934]: I1002 10:21:50.714851 4934 scope.go:117] "RemoveContainer" containerID="bbb588aba6c3c92d7afe940b0ad573ba0470987c0703f25854c72a727bebe85a" Oct 02 10:21:50 crc kubenswrapper[4934]: E1002 10:21:50.715270 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbb588aba6c3c92d7afe940b0ad573ba0470987c0703f25854c72a727bebe85a\": container with ID starting with bbb588aba6c3c92d7afe940b0ad573ba0470987c0703f25854c72a727bebe85a not found: ID does not exist" containerID="bbb588aba6c3c92d7afe940b0ad573ba0470987c0703f25854c72a727bebe85a" Oct 02 10:21:50 crc kubenswrapper[4934]: I1002 10:21:50.715324 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbb588aba6c3c92d7afe940b0ad573ba0470987c0703f25854c72a727bebe85a"} err="failed to get container status \"bbb588aba6c3c92d7afe940b0ad573ba0470987c0703f25854c72a727bebe85a\": rpc error: code = NotFound desc = could not find container \"bbb588aba6c3c92d7afe940b0ad573ba0470987c0703f25854c72a727bebe85a\": container with ID starting with bbb588aba6c3c92d7afe940b0ad573ba0470987c0703f25854c72a727bebe85a not found: ID does not exist" Oct 02 10:21:50 crc kubenswrapper[4934]: I1002 10:21:50.715366 4934 scope.go:117] "RemoveContainer" containerID="4654c94587cd3a829f74702e7d74bdf3e8598f39c12033775744aed77ecc155b" Oct 02 10:21:50 crc kubenswrapper[4934]: E1002 10:21:50.715764 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4654c94587cd3a829f74702e7d74bdf3e8598f39c12033775744aed77ecc155b\": container with ID starting with 4654c94587cd3a829f74702e7d74bdf3e8598f39c12033775744aed77ecc155b not found: ID does not exist" containerID="4654c94587cd3a829f74702e7d74bdf3e8598f39c12033775744aed77ecc155b" Oct 02 10:21:50 crc kubenswrapper[4934]: I1002 10:21:50.715810 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4654c94587cd3a829f74702e7d74bdf3e8598f39c12033775744aed77ecc155b"} err="failed to get container status \"4654c94587cd3a829f74702e7d74bdf3e8598f39c12033775744aed77ecc155b\": rpc error: code = NotFound desc = could not find container \"4654c94587cd3a829f74702e7d74bdf3e8598f39c12033775744aed77ecc155b\": container with ID starting with 4654c94587cd3a829f74702e7d74bdf3e8598f39c12033775744aed77ecc155b not found: ID does not exist" Oct 02 10:21:50 crc kubenswrapper[4934]: I1002 10:21:50.925907 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38" path="/var/lib/kubelet/pods/b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38/volumes" Oct 02 10:22:08 crc kubenswrapper[4934]: I1002 10:22:08.439324 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:22:08 crc kubenswrapper[4934]: I1002 10:22:08.439780 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:22:38 crc kubenswrapper[4934]: I1002 10:22:38.439675 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:22:38 crc kubenswrapper[4934]: I1002 10:22:38.440362 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:22:38 crc kubenswrapper[4934]: I1002 10:22:38.440445 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 10:22:38 crc kubenswrapper[4934]: I1002 10:22:38.441849 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ff97f60f399f8a24942b018acf363427f36922417bcfd215f609a25976781529"} pod="openshift-machine-config-operator/machine-config-daemon-djh5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 10:22:38 crc kubenswrapper[4934]: I1002 10:22:38.441956 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" containerID="cri-o://ff97f60f399f8a24942b018acf363427f36922417bcfd215f609a25976781529" gracePeriod=600 Oct 02 10:22:39 crc kubenswrapper[4934]: I1002 10:22:39.058992 4934 generic.go:334] "Generic (PLEG): container finished" podID="71db06ef-05b0-4f58-b251-b27117a8500a" containerID="ff97f60f399f8a24942b018acf363427f36922417bcfd215f609a25976781529" exitCode=0 Oct 02 10:22:39 crc kubenswrapper[4934]: I1002 10:22:39.059103 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerDied","Data":"ff97f60f399f8a24942b018acf363427f36922417bcfd215f609a25976781529"} Oct 02 10:22:39 crc kubenswrapper[4934]: I1002 10:22:39.060613 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerStarted","Data":"7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245"} Oct 02 10:22:39 crc kubenswrapper[4934]: I1002 10:22:39.060654 4934 scope.go:117] "RemoveContainer" containerID="f8e5c5a69906d34c263255f6b15e13b6750785cca70b5a26a248974b0b1731b1" Oct 02 10:23:50 crc kubenswrapper[4934]: I1002 10:23:50.897306 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dtzbd"] Oct 02 10:23:50 crc kubenswrapper[4934]: E1002 10:23:50.898302 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38" containerName="extract-content" Oct 02 10:23:50 crc kubenswrapper[4934]: I1002 10:23:50.898322 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38" containerName="extract-content" Oct 02 10:23:50 crc kubenswrapper[4934]: E1002 10:23:50.898350 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38" containerName="extract-utilities" Oct 02 10:23:50 crc kubenswrapper[4934]: I1002 10:23:50.898361 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38" containerName="extract-utilities" Oct 02 10:23:50 crc kubenswrapper[4934]: E1002 10:23:50.898376 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38" containerName="registry-server" Oct 02 10:23:50 crc kubenswrapper[4934]: I1002 10:23:50.898389 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38" containerName="registry-server" Oct 02 10:23:50 crc kubenswrapper[4934]: I1002 10:23:50.898651 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1f263cf-6bcc-4d6e-a1f1-f20cf9204c38" containerName="registry-server" Oct 02 10:23:50 crc kubenswrapper[4934]: I1002 10:23:50.900361 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dtzbd" Oct 02 10:23:50 crc kubenswrapper[4934]: I1002 10:23:50.923493 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dtzbd"] Oct 02 10:23:51 crc kubenswrapper[4934]: I1002 10:23:51.002281 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af81e50c-1777-4c0b-9f64-f82bfae27f1c-utilities\") pod \"redhat-marketplace-dtzbd\" (UID: \"af81e50c-1777-4c0b-9f64-f82bfae27f1c\") " pod="openshift-marketplace/redhat-marketplace-dtzbd" Oct 02 10:23:51 crc kubenswrapper[4934]: I1002 10:23:51.002439 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncx2w\" (UniqueName: \"kubernetes.io/projected/af81e50c-1777-4c0b-9f64-f82bfae27f1c-kube-api-access-ncx2w\") pod \"redhat-marketplace-dtzbd\" (UID: \"af81e50c-1777-4c0b-9f64-f82bfae27f1c\") " pod="openshift-marketplace/redhat-marketplace-dtzbd" Oct 02 10:23:51 crc kubenswrapper[4934]: I1002 10:23:51.002516 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af81e50c-1777-4c0b-9f64-f82bfae27f1c-catalog-content\") pod \"redhat-marketplace-dtzbd\" (UID: \"af81e50c-1777-4c0b-9f64-f82bfae27f1c\") " pod="openshift-marketplace/redhat-marketplace-dtzbd" Oct 02 10:23:51 crc kubenswrapper[4934]: I1002 10:23:51.103681 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af81e50c-1777-4c0b-9f64-f82bfae27f1c-catalog-content\") pod \"redhat-marketplace-dtzbd\" (UID: \"af81e50c-1777-4c0b-9f64-f82bfae27f1c\") " pod="openshift-marketplace/redhat-marketplace-dtzbd" Oct 02 10:23:51 crc kubenswrapper[4934]: I1002 10:23:51.103842 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af81e50c-1777-4c0b-9f64-f82bfae27f1c-utilities\") pod \"redhat-marketplace-dtzbd\" (UID: \"af81e50c-1777-4c0b-9f64-f82bfae27f1c\") " pod="openshift-marketplace/redhat-marketplace-dtzbd" Oct 02 10:23:51 crc kubenswrapper[4934]: I1002 10:23:51.103971 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncx2w\" (UniqueName: \"kubernetes.io/projected/af81e50c-1777-4c0b-9f64-f82bfae27f1c-kube-api-access-ncx2w\") pod \"redhat-marketplace-dtzbd\" (UID: \"af81e50c-1777-4c0b-9f64-f82bfae27f1c\") " pod="openshift-marketplace/redhat-marketplace-dtzbd" Oct 02 10:23:51 crc kubenswrapper[4934]: I1002 10:23:51.104940 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af81e50c-1777-4c0b-9f64-f82bfae27f1c-catalog-content\") pod \"redhat-marketplace-dtzbd\" (UID: \"af81e50c-1777-4c0b-9f64-f82bfae27f1c\") " pod="openshift-marketplace/redhat-marketplace-dtzbd" Oct 02 10:23:51 crc kubenswrapper[4934]: I1002 10:23:51.104975 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af81e50c-1777-4c0b-9f64-f82bfae27f1c-utilities\") pod \"redhat-marketplace-dtzbd\" (UID: \"af81e50c-1777-4c0b-9f64-f82bfae27f1c\") " pod="openshift-marketplace/redhat-marketplace-dtzbd" Oct 02 10:23:51 crc kubenswrapper[4934]: I1002 10:23:51.128424 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncx2w\" (UniqueName: \"kubernetes.io/projected/af81e50c-1777-4c0b-9f64-f82bfae27f1c-kube-api-access-ncx2w\") pod \"redhat-marketplace-dtzbd\" (UID: \"af81e50c-1777-4c0b-9f64-f82bfae27f1c\") " pod="openshift-marketplace/redhat-marketplace-dtzbd" Oct 02 10:23:51 crc kubenswrapper[4934]: I1002 10:23:51.229746 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dtzbd" Oct 02 10:23:51 crc kubenswrapper[4934]: I1002 10:23:51.717409 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dtzbd"] Oct 02 10:23:52 crc kubenswrapper[4934]: I1002 10:23:52.758235 4934 generic.go:334] "Generic (PLEG): container finished" podID="af81e50c-1777-4c0b-9f64-f82bfae27f1c" containerID="0c1256322a22f2442c06e78626b678466de6ed4e951d49d1245ec858464e7d74" exitCode=0 Oct 02 10:23:52 crc kubenswrapper[4934]: I1002 10:23:52.758366 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtzbd" event={"ID":"af81e50c-1777-4c0b-9f64-f82bfae27f1c","Type":"ContainerDied","Data":"0c1256322a22f2442c06e78626b678466de6ed4e951d49d1245ec858464e7d74"} Oct 02 10:23:52 crc kubenswrapper[4934]: I1002 10:23:52.758619 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtzbd" event={"ID":"af81e50c-1777-4c0b-9f64-f82bfae27f1c","Type":"ContainerStarted","Data":"0e9d7a8dad9f1b07dbde9040def658d717af8c2d70e938a0e72b4f3fce49bb70"} Oct 02 10:23:53 crc kubenswrapper[4934]: I1002 10:23:53.769774 4934 generic.go:334] "Generic (PLEG): container finished" podID="af81e50c-1777-4c0b-9f64-f82bfae27f1c" containerID="be8b4898aff2bee238444b5eea0b3b57840024409add130ba4999ad5caea8f09" exitCode=0 Oct 02 10:23:53 crc kubenswrapper[4934]: I1002 10:23:53.769846 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtzbd" event={"ID":"af81e50c-1777-4c0b-9f64-f82bfae27f1c","Type":"ContainerDied","Data":"be8b4898aff2bee238444b5eea0b3b57840024409add130ba4999ad5caea8f09"} Oct 02 10:23:54 crc kubenswrapper[4934]: I1002 10:23:54.781353 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtzbd" event={"ID":"af81e50c-1777-4c0b-9f64-f82bfae27f1c","Type":"ContainerStarted","Data":"a73c7787e9e8c1ac2df496722fdcbdd36223966dc50f43e806632e6574b60832"} Oct 02 10:23:54 crc kubenswrapper[4934]: I1002 10:23:54.806008 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dtzbd" podStartSLOduration=3.235686297 podStartE2EDuration="4.805988458s" podCreationTimestamp="2025-10-02 10:23:50 +0000 UTC" firstStartedPulling="2025-10-02 10:23:52.760501431 +0000 UTC m=+2104.513142993" lastFinishedPulling="2025-10-02 10:23:54.330803602 +0000 UTC m=+2106.083445154" observedRunningTime="2025-10-02 10:23:54.80317117 +0000 UTC m=+2106.555812762" watchObservedRunningTime="2025-10-02 10:23:54.805988458 +0000 UTC m=+2106.558629980" Oct 02 10:24:01 crc kubenswrapper[4934]: I1002 10:24:01.230941 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dtzbd" Oct 02 10:24:01 crc kubenswrapper[4934]: I1002 10:24:01.231821 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dtzbd" Oct 02 10:24:01 crc kubenswrapper[4934]: I1002 10:24:01.287233 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dtzbd" Oct 02 10:24:01 crc kubenswrapper[4934]: I1002 10:24:01.889232 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dtzbd" Oct 02 10:24:01 crc kubenswrapper[4934]: I1002 10:24:01.951660 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dtzbd"] Oct 02 10:24:03 crc kubenswrapper[4934]: I1002 10:24:03.856893 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dtzbd" podUID="af81e50c-1777-4c0b-9f64-f82bfae27f1c" containerName="registry-server" containerID="cri-o://a73c7787e9e8c1ac2df496722fdcbdd36223966dc50f43e806632e6574b60832" gracePeriod=2 Oct 02 10:24:04 crc kubenswrapper[4934]: I1002 10:24:04.331723 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dtzbd" Oct 02 10:24:04 crc kubenswrapper[4934]: I1002 10:24:04.519662 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncx2w\" (UniqueName: \"kubernetes.io/projected/af81e50c-1777-4c0b-9f64-f82bfae27f1c-kube-api-access-ncx2w\") pod \"af81e50c-1777-4c0b-9f64-f82bfae27f1c\" (UID: \"af81e50c-1777-4c0b-9f64-f82bfae27f1c\") " Oct 02 10:24:04 crc kubenswrapper[4934]: I1002 10:24:04.519735 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af81e50c-1777-4c0b-9f64-f82bfae27f1c-catalog-content\") pod \"af81e50c-1777-4c0b-9f64-f82bfae27f1c\" (UID: \"af81e50c-1777-4c0b-9f64-f82bfae27f1c\") " Oct 02 10:24:04 crc kubenswrapper[4934]: I1002 10:24:04.519813 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af81e50c-1777-4c0b-9f64-f82bfae27f1c-utilities\") pod \"af81e50c-1777-4c0b-9f64-f82bfae27f1c\" (UID: \"af81e50c-1777-4c0b-9f64-f82bfae27f1c\") " Oct 02 10:24:04 crc kubenswrapper[4934]: I1002 10:24:04.520816 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af81e50c-1777-4c0b-9f64-f82bfae27f1c-utilities" (OuterVolumeSpecName: "utilities") pod "af81e50c-1777-4c0b-9f64-f82bfae27f1c" (UID: "af81e50c-1777-4c0b-9f64-f82bfae27f1c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:24:04 crc kubenswrapper[4934]: I1002 10:24:04.538035 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af81e50c-1777-4c0b-9f64-f82bfae27f1c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "af81e50c-1777-4c0b-9f64-f82bfae27f1c" (UID: "af81e50c-1777-4c0b-9f64-f82bfae27f1c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:24:04 crc kubenswrapper[4934]: I1002 10:24:04.543790 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af81e50c-1777-4c0b-9f64-f82bfae27f1c-kube-api-access-ncx2w" (OuterVolumeSpecName: "kube-api-access-ncx2w") pod "af81e50c-1777-4c0b-9f64-f82bfae27f1c" (UID: "af81e50c-1777-4c0b-9f64-f82bfae27f1c"). InnerVolumeSpecName "kube-api-access-ncx2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:24:04 crc kubenswrapper[4934]: I1002 10:24:04.621401 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncx2w\" (UniqueName: \"kubernetes.io/projected/af81e50c-1777-4c0b-9f64-f82bfae27f1c-kube-api-access-ncx2w\") on node \"crc\" DevicePath \"\"" Oct 02 10:24:04 crc kubenswrapper[4934]: I1002 10:24:04.621454 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af81e50c-1777-4c0b-9f64-f82bfae27f1c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:24:04 crc kubenswrapper[4934]: I1002 10:24:04.621469 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af81e50c-1777-4c0b-9f64-f82bfae27f1c-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:24:04 crc kubenswrapper[4934]: I1002 10:24:04.869261 4934 generic.go:334] "Generic (PLEG): container finished" podID="af81e50c-1777-4c0b-9f64-f82bfae27f1c" containerID="a73c7787e9e8c1ac2df496722fdcbdd36223966dc50f43e806632e6574b60832" exitCode=0 Oct 02 10:24:04 crc kubenswrapper[4934]: I1002 10:24:04.869304 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtzbd" event={"ID":"af81e50c-1777-4c0b-9f64-f82bfae27f1c","Type":"ContainerDied","Data":"a73c7787e9e8c1ac2df496722fdcbdd36223966dc50f43e806632e6574b60832"} Oct 02 10:24:04 crc kubenswrapper[4934]: I1002 10:24:04.869332 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtzbd" event={"ID":"af81e50c-1777-4c0b-9f64-f82bfae27f1c","Type":"ContainerDied","Data":"0e9d7a8dad9f1b07dbde9040def658d717af8c2d70e938a0e72b4f3fce49bb70"} Oct 02 10:24:04 crc kubenswrapper[4934]: I1002 10:24:04.869349 4934 scope.go:117] "RemoveContainer" containerID="a73c7787e9e8c1ac2df496722fdcbdd36223966dc50f43e806632e6574b60832" Oct 02 10:24:04 crc kubenswrapper[4934]: I1002 10:24:04.869391 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dtzbd" Oct 02 10:24:04 crc kubenswrapper[4934]: I1002 10:24:04.892448 4934 scope.go:117] "RemoveContainer" containerID="be8b4898aff2bee238444b5eea0b3b57840024409add130ba4999ad5caea8f09" Oct 02 10:24:04 crc kubenswrapper[4934]: I1002 10:24:04.943351 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dtzbd"] Oct 02 10:24:04 crc kubenswrapper[4934]: I1002 10:24:04.951844 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dtzbd"] Oct 02 10:24:04 crc kubenswrapper[4934]: I1002 10:24:04.956757 4934 scope.go:117] "RemoveContainer" containerID="0c1256322a22f2442c06e78626b678466de6ed4e951d49d1245ec858464e7d74" Oct 02 10:24:04 crc kubenswrapper[4934]: I1002 10:24:04.977996 4934 scope.go:117] "RemoveContainer" containerID="a73c7787e9e8c1ac2df496722fdcbdd36223966dc50f43e806632e6574b60832" Oct 02 10:24:04 crc kubenswrapper[4934]: E1002 10:24:04.978391 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a73c7787e9e8c1ac2df496722fdcbdd36223966dc50f43e806632e6574b60832\": container with ID starting with a73c7787e9e8c1ac2df496722fdcbdd36223966dc50f43e806632e6574b60832 not found: ID does not exist" containerID="a73c7787e9e8c1ac2df496722fdcbdd36223966dc50f43e806632e6574b60832" Oct 02 10:24:04 crc kubenswrapper[4934]: I1002 10:24:04.978419 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a73c7787e9e8c1ac2df496722fdcbdd36223966dc50f43e806632e6574b60832"} err="failed to get container status \"a73c7787e9e8c1ac2df496722fdcbdd36223966dc50f43e806632e6574b60832\": rpc error: code = NotFound desc = could not find container \"a73c7787e9e8c1ac2df496722fdcbdd36223966dc50f43e806632e6574b60832\": container with ID starting with a73c7787e9e8c1ac2df496722fdcbdd36223966dc50f43e806632e6574b60832 not found: ID does not exist" Oct 02 10:24:04 crc kubenswrapper[4934]: I1002 10:24:04.978441 4934 scope.go:117] "RemoveContainer" containerID="be8b4898aff2bee238444b5eea0b3b57840024409add130ba4999ad5caea8f09" Oct 02 10:24:04 crc kubenswrapper[4934]: E1002 10:24:04.978755 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be8b4898aff2bee238444b5eea0b3b57840024409add130ba4999ad5caea8f09\": container with ID starting with be8b4898aff2bee238444b5eea0b3b57840024409add130ba4999ad5caea8f09 not found: ID does not exist" containerID="be8b4898aff2bee238444b5eea0b3b57840024409add130ba4999ad5caea8f09" Oct 02 10:24:04 crc kubenswrapper[4934]: I1002 10:24:04.978779 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be8b4898aff2bee238444b5eea0b3b57840024409add130ba4999ad5caea8f09"} err="failed to get container status \"be8b4898aff2bee238444b5eea0b3b57840024409add130ba4999ad5caea8f09\": rpc error: code = NotFound desc = could not find container \"be8b4898aff2bee238444b5eea0b3b57840024409add130ba4999ad5caea8f09\": container with ID starting with be8b4898aff2bee238444b5eea0b3b57840024409add130ba4999ad5caea8f09 not found: ID does not exist" Oct 02 10:24:04 crc kubenswrapper[4934]: I1002 10:24:04.978796 4934 scope.go:117] "RemoveContainer" containerID="0c1256322a22f2442c06e78626b678466de6ed4e951d49d1245ec858464e7d74" Oct 02 10:24:04 crc kubenswrapper[4934]: E1002 10:24:04.979052 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c1256322a22f2442c06e78626b678466de6ed4e951d49d1245ec858464e7d74\": container with ID starting with 0c1256322a22f2442c06e78626b678466de6ed4e951d49d1245ec858464e7d74 not found: ID does not exist" containerID="0c1256322a22f2442c06e78626b678466de6ed4e951d49d1245ec858464e7d74" Oct 02 10:24:04 crc kubenswrapper[4934]: I1002 10:24:04.979069 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c1256322a22f2442c06e78626b678466de6ed4e951d49d1245ec858464e7d74"} err="failed to get container status \"0c1256322a22f2442c06e78626b678466de6ed4e951d49d1245ec858464e7d74\": rpc error: code = NotFound desc = could not find container \"0c1256322a22f2442c06e78626b678466de6ed4e951d49d1245ec858464e7d74\": container with ID starting with 0c1256322a22f2442c06e78626b678466de6ed4e951d49d1245ec858464e7d74 not found: ID does not exist" Oct 02 10:24:06 crc kubenswrapper[4934]: I1002 10:24:06.931178 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af81e50c-1777-4c0b-9f64-f82bfae27f1c" path="/var/lib/kubelet/pods/af81e50c-1777-4c0b-9f64-f82bfae27f1c/volumes" Oct 02 10:24:38 crc kubenswrapper[4934]: I1002 10:24:38.440899 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:24:38 crc kubenswrapper[4934]: I1002 10:24:38.441645 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:25:01 crc kubenswrapper[4934]: I1002 10:25:01.168384 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5lwhm"] Oct 02 10:25:01 crc kubenswrapper[4934]: E1002 10:25:01.169491 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af81e50c-1777-4c0b-9f64-f82bfae27f1c" containerName="extract-content" Oct 02 10:25:01 crc kubenswrapper[4934]: I1002 10:25:01.169512 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="af81e50c-1777-4c0b-9f64-f82bfae27f1c" containerName="extract-content" Oct 02 10:25:01 crc kubenswrapper[4934]: E1002 10:25:01.169533 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af81e50c-1777-4c0b-9f64-f82bfae27f1c" containerName="registry-server" Oct 02 10:25:01 crc kubenswrapper[4934]: I1002 10:25:01.169548 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="af81e50c-1777-4c0b-9f64-f82bfae27f1c" containerName="registry-server" Oct 02 10:25:01 crc kubenswrapper[4934]: E1002 10:25:01.169609 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af81e50c-1777-4c0b-9f64-f82bfae27f1c" containerName="extract-utilities" Oct 02 10:25:01 crc kubenswrapper[4934]: I1002 10:25:01.169623 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="af81e50c-1777-4c0b-9f64-f82bfae27f1c" containerName="extract-utilities" Oct 02 10:25:01 crc kubenswrapper[4934]: I1002 10:25:01.169872 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="af81e50c-1777-4c0b-9f64-f82bfae27f1c" containerName="registry-server" Oct 02 10:25:01 crc kubenswrapper[4934]: I1002 10:25:01.171662 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5lwhm" Oct 02 10:25:01 crc kubenswrapper[4934]: I1002 10:25:01.188377 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5lwhm"] Oct 02 10:25:01 crc kubenswrapper[4934]: I1002 10:25:01.324794 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f8x6\" (UniqueName: \"kubernetes.io/projected/c3b01970-7896-4bd6-a9ab-794cd0e7a2fa-kube-api-access-2f8x6\") pod \"certified-operators-5lwhm\" (UID: \"c3b01970-7896-4bd6-a9ab-794cd0e7a2fa\") " pod="openshift-marketplace/certified-operators-5lwhm" Oct 02 10:25:01 crc kubenswrapper[4934]: I1002 10:25:01.325146 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3b01970-7896-4bd6-a9ab-794cd0e7a2fa-catalog-content\") pod \"certified-operators-5lwhm\" (UID: \"c3b01970-7896-4bd6-a9ab-794cd0e7a2fa\") " pod="openshift-marketplace/certified-operators-5lwhm" Oct 02 10:25:01 crc kubenswrapper[4934]: I1002 10:25:01.325314 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3b01970-7896-4bd6-a9ab-794cd0e7a2fa-utilities\") pod \"certified-operators-5lwhm\" (UID: \"c3b01970-7896-4bd6-a9ab-794cd0e7a2fa\") " pod="openshift-marketplace/certified-operators-5lwhm" Oct 02 10:25:01 crc kubenswrapper[4934]: I1002 10:25:01.426170 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3b01970-7896-4bd6-a9ab-794cd0e7a2fa-catalog-content\") pod \"certified-operators-5lwhm\" (UID: \"c3b01970-7896-4bd6-a9ab-794cd0e7a2fa\") " pod="openshift-marketplace/certified-operators-5lwhm" Oct 02 10:25:01 crc kubenswrapper[4934]: I1002 10:25:01.426232 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3b01970-7896-4bd6-a9ab-794cd0e7a2fa-utilities\") pod \"certified-operators-5lwhm\" (UID: \"c3b01970-7896-4bd6-a9ab-794cd0e7a2fa\") " pod="openshift-marketplace/certified-operators-5lwhm" Oct 02 10:25:01 crc kubenswrapper[4934]: I1002 10:25:01.426300 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2f8x6\" (UniqueName: \"kubernetes.io/projected/c3b01970-7896-4bd6-a9ab-794cd0e7a2fa-kube-api-access-2f8x6\") pod \"certified-operators-5lwhm\" (UID: \"c3b01970-7896-4bd6-a9ab-794cd0e7a2fa\") " pod="openshift-marketplace/certified-operators-5lwhm" Oct 02 10:25:01 crc kubenswrapper[4934]: I1002 10:25:01.427139 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3b01970-7896-4bd6-a9ab-794cd0e7a2fa-catalog-content\") pod \"certified-operators-5lwhm\" (UID: \"c3b01970-7896-4bd6-a9ab-794cd0e7a2fa\") " pod="openshift-marketplace/certified-operators-5lwhm" Oct 02 10:25:01 crc kubenswrapper[4934]: I1002 10:25:01.427298 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3b01970-7896-4bd6-a9ab-794cd0e7a2fa-utilities\") pod \"certified-operators-5lwhm\" (UID: \"c3b01970-7896-4bd6-a9ab-794cd0e7a2fa\") " pod="openshift-marketplace/certified-operators-5lwhm" Oct 02 10:25:01 crc kubenswrapper[4934]: I1002 10:25:01.453614 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2f8x6\" (UniqueName: \"kubernetes.io/projected/c3b01970-7896-4bd6-a9ab-794cd0e7a2fa-kube-api-access-2f8x6\") pod \"certified-operators-5lwhm\" (UID: \"c3b01970-7896-4bd6-a9ab-794cd0e7a2fa\") " pod="openshift-marketplace/certified-operators-5lwhm" Oct 02 10:25:01 crc kubenswrapper[4934]: I1002 10:25:01.501379 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5lwhm" Oct 02 10:25:02 crc kubenswrapper[4934]: I1002 10:25:02.051253 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5lwhm"] Oct 02 10:25:02 crc kubenswrapper[4934]: I1002 10:25:02.435340 4934 generic.go:334] "Generic (PLEG): container finished" podID="c3b01970-7896-4bd6-a9ab-794cd0e7a2fa" containerID="8c2420368b340bec8aee5e107494ca6235092e5b2267014039c4253af2f42282" exitCode=0 Oct 02 10:25:02 crc kubenswrapper[4934]: I1002 10:25:02.435549 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5lwhm" event={"ID":"c3b01970-7896-4bd6-a9ab-794cd0e7a2fa","Type":"ContainerDied","Data":"8c2420368b340bec8aee5e107494ca6235092e5b2267014039c4253af2f42282"} Oct 02 10:25:02 crc kubenswrapper[4934]: I1002 10:25:02.435704 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5lwhm" event={"ID":"c3b01970-7896-4bd6-a9ab-794cd0e7a2fa","Type":"ContainerStarted","Data":"a52b5e3c17714be9e20a18dfb35a3b843a29602b9f339b73ca1ec1dc6434dbf5"} Oct 02 10:25:03 crc kubenswrapper[4934]: I1002 10:25:03.449334 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5lwhm" event={"ID":"c3b01970-7896-4bd6-a9ab-794cd0e7a2fa","Type":"ContainerStarted","Data":"df64ed92b1a56471f4f51b50c3c01ea724d155bc2c868ae7d5dd7fb5cd73b0f2"} Oct 02 10:25:04 crc kubenswrapper[4934]: I1002 10:25:04.462702 4934 generic.go:334] "Generic (PLEG): container finished" podID="c3b01970-7896-4bd6-a9ab-794cd0e7a2fa" containerID="df64ed92b1a56471f4f51b50c3c01ea724d155bc2c868ae7d5dd7fb5cd73b0f2" exitCode=0 Oct 02 10:25:04 crc kubenswrapper[4934]: I1002 10:25:04.463055 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5lwhm" event={"ID":"c3b01970-7896-4bd6-a9ab-794cd0e7a2fa","Type":"ContainerDied","Data":"df64ed92b1a56471f4f51b50c3c01ea724d155bc2c868ae7d5dd7fb5cd73b0f2"} Oct 02 10:25:05 crc kubenswrapper[4934]: I1002 10:25:05.478275 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5lwhm" event={"ID":"c3b01970-7896-4bd6-a9ab-794cd0e7a2fa","Type":"ContainerStarted","Data":"35fb742843275f096bb5125aa319803c25c3f7856e31d0c1afc535abc114b7f5"} Oct 02 10:25:05 crc kubenswrapper[4934]: I1002 10:25:05.519779 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5lwhm" podStartSLOduration=2.055711495 podStartE2EDuration="4.519749596s" podCreationTimestamp="2025-10-02 10:25:01 +0000 UTC" firstStartedPulling="2025-10-02 10:25:02.439645526 +0000 UTC m=+2174.192287048" lastFinishedPulling="2025-10-02 10:25:04.903683587 +0000 UTC m=+2176.656325149" observedRunningTime="2025-10-02 10:25:05.509922102 +0000 UTC m=+2177.262563664" watchObservedRunningTime="2025-10-02 10:25:05.519749596 +0000 UTC m=+2177.272391158" Oct 02 10:25:08 crc kubenswrapper[4934]: I1002 10:25:08.439760 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:25:08 crc kubenswrapper[4934]: I1002 10:25:08.440232 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:25:11 crc kubenswrapper[4934]: I1002 10:25:11.501950 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5lwhm" Oct 02 10:25:11 crc kubenswrapper[4934]: I1002 10:25:11.502482 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5lwhm" Oct 02 10:25:11 crc kubenswrapper[4934]: I1002 10:25:11.585712 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5lwhm" Oct 02 10:25:11 crc kubenswrapper[4934]: I1002 10:25:11.659288 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5lwhm" Oct 02 10:25:11 crc kubenswrapper[4934]: I1002 10:25:11.836834 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5lwhm"] Oct 02 10:25:13 crc kubenswrapper[4934]: I1002 10:25:13.549850 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5lwhm" podUID="c3b01970-7896-4bd6-a9ab-794cd0e7a2fa" containerName="registry-server" containerID="cri-o://35fb742843275f096bb5125aa319803c25c3f7856e31d0c1afc535abc114b7f5" gracePeriod=2 Oct 02 10:25:14 crc kubenswrapper[4934]: I1002 10:25:14.054796 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5lwhm" Oct 02 10:25:14 crc kubenswrapper[4934]: I1002 10:25:14.069625 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3b01970-7896-4bd6-a9ab-794cd0e7a2fa-utilities\") pod \"c3b01970-7896-4bd6-a9ab-794cd0e7a2fa\" (UID: \"c3b01970-7896-4bd6-a9ab-794cd0e7a2fa\") " Oct 02 10:25:14 crc kubenswrapper[4934]: I1002 10:25:14.069904 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2f8x6\" (UniqueName: \"kubernetes.io/projected/c3b01970-7896-4bd6-a9ab-794cd0e7a2fa-kube-api-access-2f8x6\") pod \"c3b01970-7896-4bd6-a9ab-794cd0e7a2fa\" (UID: \"c3b01970-7896-4bd6-a9ab-794cd0e7a2fa\") " Oct 02 10:25:14 crc kubenswrapper[4934]: I1002 10:25:14.070867 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3b01970-7896-4bd6-a9ab-794cd0e7a2fa-catalog-content\") pod \"c3b01970-7896-4bd6-a9ab-794cd0e7a2fa\" (UID: \"c3b01970-7896-4bd6-a9ab-794cd0e7a2fa\") " Oct 02 10:25:14 crc kubenswrapper[4934]: I1002 10:25:14.072966 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3b01970-7896-4bd6-a9ab-794cd0e7a2fa-utilities" (OuterVolumeSpecName: "utilities") pod "c3b01970-7896-4bd6-a9ab-794cd0e7a2fa" (UID: "c3b01970-7896-4bd6-a9ab-794cd0e7a2fa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:25:14 crc kubenswrapper[4934]: I1002 10:25:14.082216 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3b01970-7896-4bd6-a9ab-794cd0e7a2fa-kube-api-access-2f8x6" (OuterVolumeSpecName: "kube-api-access-2f8x6") pod "c3b01970-7896-4bd6-a9ab-794cd0e7a2fa" (UID: "c3b01970-7896-4bd6-a9ab-794cd0e7a2fa"). InnerVolumeSpecName "kube-api-access-2f8x6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:25:14 crc kubenswrapper[4934]: I1002 10:25:14.164895 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3b01970-7896-4bd6-a9ab-794cd0e7a2fa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c3b01970-7896-4bd6-a9ab-794cd0e7a2fa" (UID: "c3b01970-7896-4bd6-a9ab-794cd0e7a2fa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:25:14 crc kubenswrapper[4934]: I1002 10:25:14.172445 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3b01970-7896-4bd6-a9ab-794cd0e7a2fa-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:25:14 crc kubenswrapper[4934]: I1002 10:25:14.172472 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3b01970-7896-4bd6-a9ab-794cd0e7a2fa-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:25:14 crc kubenswrapper[4934]: I1002 10:25:14.172481 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2f8x6\" (UniqueName: \"kubernetes.io/projected/c3b01970-7896-4bd6-a9ab-794cd0e7a2fa-kube-api-access-2f8x6\") on node \"crc\" DevicePath \"\"" Oct 02 10:25:14 crc kubenswrapper[4934]: I1002 10:25:14.564243 4934 generic.go:334] "Generic (PLEG): container finished" podID="c3b01970-7896-4bd6-a9ab-794cd0e7a2fa" containerID="35fb742843275f096bb5125aa319803c25c3f7856e31d0c1afc535abc114b7f5" exitCode=0 Oct 02 10:25:14 crc kubenswrapper[4934]: I1002 10:25:14.564326 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5lwhm" Oct 02 10:25:14 crc kubenswrapper[4934]: I1002 10:25:14.564339 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5lwhm" event={"ID":"c3b01970-7896-4bd6-a9ab-794cd0e7a2fa","Type":"ContainerDied","Data":"35fb742843275f096bb5125aa319803c25c3f7856e31d0c1afc535abc114b7f5"} Oct 02 10:25:14 crc kubenswrapper[4934]: I1002 10:25:14.564448 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5lwhm" event={"ID":"c3b01970-7896-4bd6-a9ab-794cd0e7a2fa","Type":"ContainerDied","Data":"a52b5e3c17714be9e20a18dfb35a3b843a29602b9f339b73ca1ec1dc6434dbf5"} Oct 02 10:25:14 crc kubenswrapper[4934]: I1002 10:25:14.564520 4934 scope.go:117] "RemoveContainer" containerID="35fb742843275f096bb5125aa319803c25c3f7856e31d0c1afc535abc114b7f5" Oct 02 10:25:14 crc kubenswrapper[4934]: I1002 10:25:14.600454 4934 scope.go:117] "RemoveContainer" containerID="df64ed92b1a56471f4f51b50c3c01ea724d155bc2c868ae7d5dd7fb5cd73b0f2" Oct 02 10:25:14 crc kubenswrapper[4934]: I1002 10:25:14.632710 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5lwhm"] Oct 02 10:25:14 crc kubenswrapper[4934]: I1002 10:25:14.641228 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5lwhm"] Oct 02 10:25:14 crc kubenswrapper[4934]: I1002 10:25:14.647479 4934 scope.go:117] "RemoveContainer" containerID="8c2420368b340bec8aee5e107494ca6235092e5b2267014039c4253af2f42282" Oct 02 10:25:14 crc kubenswrapper[4934]: I1002 10:25:14.681605 4934 scope.go:117] "RemoveContainer" containerID="35fb742843275f096bb5125aa319803c25c3f7856e31d0c1afc535abc114b7f5" Oct 02 10:25:14 crc kubenswrapper[4934]: E1002 10:25:14.682095 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35fb742843275f096bb5125aa319803c25c3f7856e31d0c1afc535abc114b7f5\": container with ID starting with 35fb742843275f096bb5125aa319803c25c3f7856e31d0c1afc535abc114b7f5 not found: ID does not exist" containerID="35fb742843275f096bb5125aa319803c25c3f7856e31d0c1afc535abc114b7f5" Oct 02 10:25:14 crc kubenswrapper[4934]: I1002 10:25:14.682132 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35fb742843275f096bb5125aa319803c25c3f7856e31d0c1afc535abc114b7f5"} err="failed to get container status \"35fb742843275f096bb5125aa319803c25c3f7856e31d0c1afc535abc114b7f5\": rpc error: code = NotFound desc = could not find container \"35fb742843275f096bb5125aa319803c25c3f7856e31d0c1afc535abc114b7f5\": container with ID starting with 35fb742843275f096bb5125aa319803c25c3f7856e31d0c1afc535abc114b7f5 not found: ID does not exist" Oct 02 10:25:14 crc kubenswrapper[4934]: I1002 10:25:14.682161 4934 scope.go:117] "RemoveContainer" containerID="df64ed92b1a56471f4f51b50c3c01ea724d155bc2c868ae7d5dd7fb5cd73b0f2" Oct 02 10:25:14 crc kubenswrapper[4934]: E1002 10:25:14.682659 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df64ed92b1a56471f4f51b50c3c01ea724d155bc2c868ae7d5dd7fb5cd73b0f2\": container with ID starting with df64ed92b1a56471f4f51b50c3c01ea724d155bc2c868ae7d5dd7fb5cd73b0f2 not found: ID does not exist" containerID="df64ed92b1a56471f4f51b50c3c01ea724d155bc2c868ae7d5dd7fb5cd73b0f2" Oct 02 10:25:14 crc kubenswrapper[4934]: I1002 10:25:14.682715 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df64ed92b1a56471f4f51b50c3c01ea724d155bc2c868ae7d5dd7fb5cd73b0f2"} err="failed to get container status \"df64ed92b1a56471f4f51b50c3c01ea724d155bc2c868ae7d5dd7fb5cd73b0f2\": rpc error: code = NotFound desc = could not find container \"df64ed92b1a56471f4f51b50c3c01ea724d155bc2c868ae7d5dd7fb5cd73b0f2\": container with ID starting with df64ed92b1a56471f4f51b50c3c01ea724d155bc2c868ae7d5dd7fb5cd73b0f2 not found: ID does not exist" Oct 02 10:25:14 crc kubenswrapper[4934]: I1002 10:25:14.682756 4934 scope.go:117] "RemoveContainer" containerID="8c2420368b340bec8aee5e107494ca6235092e5b2267014039c4253af2f42282" Oct 02 10:25:14 crc kubenswrapper[4934]: E1002 10:25:14.683304 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c2420368b340bec8aee5e107494ca6235092e5b2267014039c4253af2f42282\": container with ID starting with 8c2420368b340bec8aee5e107494ca6235092e5b2267014039c4253af2f42282 not found: ID does not exist" containerID="8c2420368b340bec8aee5e107494ca6235092e5b2267014039c4253af2f42282" Oct 02 10:25:14 crc kubenswrapper[4934]: I1002 10:25:14.683368 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c2420368b340bec8aee5e107494ca6235092e5b2267014039c4253af2f42282"} err="failed to get container status \"8c2420368b340bec8aee5e107494ca6235092e5b2267014039c4253af2f42282\": rpc error: code = NotFound desc = could not find container \"8c2420368b340bec8aee5e107494ca6235092e5b2267014039c4253af2f42282\": container with ID starting with 8c2420368b340bec8aee5e107494ca6235092e5b2267014039c4253af2f42282 not found: ID does not exist" Oct 02 10:25:14 crc kubenswrapper[4934]: I1002 10:25:14.928865 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3b01970-7896-4bd6-a9ab-794cd0e7a2fa" path="/var/lib/kubelet/pods/c3b01970-7896-4bd6-a9ab-794cd0e7a2fa/volumes" Oct 02 10:25:38 crc kubenswrapper[4934]: I1002 10:25:38.440106 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:25:38 crc kubenswrapper[4934]: I1002 10:25:38.440867 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:25:38 crc kubenswrapper[4934]: I1002 10:25:38.441062 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 10:25:38 crc kubenswrapper[4934]: I1002 10:25:38.442683 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245"} pod="openshift-machine-config-operator/machine-config-daemon-djh5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 10:25:38 crc kubenswrapper[4934]: I1002 10:25:38.442815 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" containerID="cri-o://7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245" gracePeriod=600 Oct 02 10:25:38 crc kubenswrapper[4934]: E1002 10:25:38.581321 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:25:38 crc kubenswrapper[4934]: I1002 10:25:38.787909 4934 generic.go:334] "Generic (PLEG): container finished" podID="71db06ef-05b0-4f58-b251-b27117a8500a" containerID="7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245" exitCode=0 Oct 02 10:25:38 crc kubenswrapper[4934]: I1002 10:25:38.787964 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerDied","Data":"7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245"} Oct 02 10:25:38 crc kubenswrapper[4934]: I1002 10:25:38.788014 4934 scope.go:117] "RemoveContainer" containerID="ff97f60f399f8a24942b018acf363427f36922417bcfd215f609a25976781529" Oct 02 10:25:38 crc kubenswrapper[4934]: I1002 10:25:38.788755 4934 scope.go:117] "RemoveContainer" containerID="7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245" Oct 02 10:25:38 crc kubenswrapper[4934]: E1002 10:25:38.790416 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:25:49 crc kubenswrapper[4934]: I1002 10:25:49.913425 4934 scope.go:117] "RemoveContainer" containerID="7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245" Oct 02 10:25:49 crc kubenswrapper[4934]: E1002 10:25:49.914725 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:26:00 crc kubenswrapper[4934]: I1002 10:26:00.913707 4934 scope.go:117] "RemoveContainer" containerID="7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245" Oct 02 10:26:00 crc kubenswrapper[4934]: E1002 10:26:00.914817 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:26:14 crc kubenswrapper[4934]: I1002 10:26:14.913934 4934 scope.go:117] "RemoveContainer" containerID="7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245" Oct 02 10:26:14 crc kubenswrapper[4934]: E1002 10:26:14.914977 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:26:25 crc kubenswrapper[4934]: I1002 10:26:25.749945 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6gpk4"] Oct 02 10:26:25 crc kubenswrapper[4934]: E1002 10:26:25.750972 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3b01970-7896-4bd6-a9ab-794cd0e7a2fa" containerName="extract-content" Oct 02 10:26:25 crc kubenswrapper[4934]: I1002 10:26:25.750993 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3b01970-7896-4bd6-a9ab-794cd0e7a2fa" containerName="extract-content" Oct 02 10:26:25 crc kubenswrapper[4934]: E1002 10:26:25.751024 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3b01970-7896-4bd6-a9ab-794cd0e7a2fa" containerName="extract-utilities" Oct 02 10:26:25 crc kubenswrapper[4934]: I1002 10:26:25.751037 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3b01970-7896-4bd6-a9ab-794cd0e7a2fa" containerName="extract-utilities" Oct 02 10:26:25 crc kubenswrapper[4934]: E1002 10:26:25.751063 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3b01970-7896-4bd6-a9ab-794cd0e7a2fa" containerName="registry-server" Oct 02 10:26:25 crc kubenswrapper[4934]: I1002 10:26:25.751077 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3b01970-7896-4bd6-a9ab-794cd0e7a2fa" containerName="registry-server" Oct 02 10:26:25 crc kubenswrapper[4934]: I1002 10:26:25.751362 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3b01970-7896-4bd6-a9ab-794cd0e7a2fa" containerName="registry-server" Oct 02 10:26:25 crc kubenswrapper[4934]: I1002 10:26:25.753378 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6gpk4" Oct 02 10:26:25 crc kubenswrapper[4934]: I1002 10:26:25.772319 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6gpk4"] Oct 02 10:26:25 crc kubenswrapper[4934]: I1002 10:26:25.855842 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnmxm\" (UniqueName: \"kubernetes.io/projected/553f29c5-eb91-4b7e-a5f3-6eddbde93b65-kube-api-access-qnmxm\") pod \"redhat-operators-6gpk4\" (UID: \"553f29c5-eb91-4b7e-a5f3-6eddbde93b65\") " pod="openshift-marketplace/redhat-operators-6gpk4" Oct 02 10:26:25 crc kubenswrapper[4934]: I1002 10:26:25.855928 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/553f29c5-eb91-4b7e-a5f3-6eddbde93b65-utilities\") pod \"redhat-operators-6gpk4\" (UID: \"553f29c5-eb91-4b7e-a5f3-6eddbde93b65\") " pod="openshift-marketplace/redhat-operators-6gpk4" Oct 02 10:26:25 crc kubenswrapper[4934]: I1002 10:26:25.855952 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/553f29c5-eb91-4b7e-a5f3-6eddbde93b65-catalog-content\") pod \"redhat-operators-6gpk4\" (UID: \"553f29c5-eb91-4b7e-a5f3-6eddbde93b65\") " pod="openshift-marketplace/redhat-operators-6gpk4" Oct 02 10:26:25 crc kubenswrapper[4934]: I1002 10:26:25.957683 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/553f29c5-eb91-4b7e-a5f3-6eddbde93b65-utilities\") pod \"redhat-operators-6gpk4\" (UID: \"553f29c5-eb91-4b7e-a5f3-6eddbde93b65\") " pod="openshift-marketplace/redhat-operators-6gpk4" Oct 02 10:26:25 crc kubenswrapper[4934]: I1002 10:26:25.957735 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/553f29c5-eb91-4b7e-a5f3-6eddbde93b65-catalog-content\") pod \"redhat-operators-6gpk4\" (UID: \"553f29c5-eb91-4b7e-a5f3-6eddbde93b65\") " pod="openshift-marketplace/redhat-operators-6gpk4" Oct 02 10:26:25 crc kubenswrapper[4934]: I1002 10:26:25.958286 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/553f29c5-eb91-4b7e-a5f3-6eddbde93b65-catalog-content\") pod \"redhat-operators-6gpk4\" (UID: \"553f29c5-eb91-4b7e-a5f3-6eddbde93b65\") " pod="openshift-marketplace/redhat-operators-6gpk4" Oct 02 10:26:25 crc kubenswrapper[4934]: I1002 10:26:25.958398 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/553f29c5-eb91-4b7e-a5f3-6eddbde93b65-utilities\") pod \"redhat-operators-6gpk4\" (UID: \"553f29c5-eb91-4b7e-a5f3-6eddbde93b65\") " pod="openshift-marketplace/redhat-operators-6gpk4" Oct 02 10:26:25 crc kubenswrapper[4934]: I1002 10:26:25.958645 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnmxm\" (UniqueName: \"kubernetes.io/projected/553f29c5-eb91-4b7e-a5f3-6eddbde93b65-kube-api-access-qnmxm\") pod \"redhat-operators-6gpk4\" (UID: \"553f29c5-eb91-4b7e-a5f3-6eddbde93b65\") " pod="openshift-marketplace/redhat-operators-6gpk4" Oct 02 10:26:25 crc kubenswrapper[4934]: I1002 10:26:25.990100 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnmxm\" (UniqueName: \"kubernetes.io/projected/553f29c5-eb91-4b7e-a5f3-6eddbde93b65-kube-api-access-qnmxm\") pod \"redhat-operators-6gpk4\" (UID: \"553f29c5-eb91-4b7e-a5f3-6eddbde93b65\") " pod="openshift-marketplace/redhat-operators-6gpk4" Oct 02 10:26:26 crc kubenswrapper[4934]: I1002 10:26:26.084751 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6gpk4" Oct 02 10:26:26 crc kubenswrapper[4934]: I1002 10:26:26.566832 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6gpk4"] Oct 02 10:26:26 crc kubenswrapper[4934]: I1002 10:26:26.912901 4934 scope.go:117] "RemoveContainer" containerID="7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245" Oct 02 10:26:26 crc kubenswrapper[4934]: E1002 10:26:26.914076 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:26:27 crc kubenswrapper[4934]: I1002 10:26:27.280908 4934 generic.go:334] "Generic (PLEG): container finished" podID="553f29c5-eb91-4b7e-a5f3-6eddbde93b65" containerID="f4a1f4c260bd46b6f84dfde602495922395e1640b4ee894a9d68d0c3acf97754" exitCode=0 Oct 02 10:26:27 crc kubenswrapper[4934]: I1002 10:26:27.280971 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6gpk4" event={"ID":"553f29c5-eb91-4b7e-a5f3-6eddbde93b65","Type":"ContainerDied","Data":"f4a1f4c260bd46b6f84dfde602495922395e1640b4ee894a9d68d0c3acf97754"} Oct 02 10:26:27 crc kubenswrapper[4934]: I1002 10:26:27.281032 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6gpk4" event={"ID":"553f29c5-eb91-4b7e-a5f3-6eddbde93b65","Type":"ContainerStarted","Data":"e2dc6d950e2ffc2dc2a39ff83b9f847ca20240d33468d3f47c39984c550c711a"} Oct 02 10:26:29 crc kubenswrapper[4934]: I1002 10:26:29.303985 4934 generic.go:334] "Generic (PLEG): container finished" podID="553f29c5-eb91-4b7e-a5f3-6eddbde93b65" containerID="b46d13a5f1a0ae5d04ad576a5aa385e70220a85a3f7cd69991b1b1db7cc7148e" exitCode=0 Oct 02 10:26:29 crc kubenswrapper[4934]: I1002 10:26:29.304055 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6gpk4" event={"ID":"553f29c5-eb91-4b7e-a5f3-6eddbde93b65","Type":"ContainerDied","Data":"b46d13a5f1a0ae5d04ad576a5aa385e70220a85a3f7cd69991b1b1db7cc7148e"} Oct 02 10:26:30 crc kubenswrapper[4934]: I1002 10:26:30.313554 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6gpk4" event={"ID":"553f29c5-eb91-4b7e-a5f3-6eddbde93b65","Type":"ContainerStarted","Data":"5062ac9255d4cb3d4d61a437e1f801a926b5f03f0f7d94eda107534d7f6b0ce7"} Oct 02 10:26:30 crc kubenswrapper[4934]: I1002 10:26:30.341869 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6gpk4" podStartSLOduration=2.804408708 podStartE2EDuration="5.341844579s" podCreationTimestamp="2025-10-02 10:26:25 +0000 UTC" firstStartedPulling="2025-10-02 10:26:27.282843945 +0000 UTC m=+2259.035485477" lastFinishedPulling="2025-10-02 10:26:29.820279786 +0000 UTC m=+2261.572921348" observedRunningTime="2025-10-02 10:26:30.338379543 +0000 UTC m=+2262.091021105" watchObservedRunningTime="2025-10-02 10:26:30.341844579 +0000 UTC m=+2262.094486141" Oct 02 10:26:36 crc kubenswrapper[4934]: I1002 10:26:36.084995 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6gpk4" Oct 02 10:26:36 crc kubenswrapper[4934]: I1002 10:26:36.085547 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6gpk4" Oct 02 10:26:36 crc kubenswrapper[4934]: I1002 10:26:36.167266 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6gpk4" Oct 02 10:26:36 crc kubenswrapper[4934]: I1002 10:26:36.422662 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6gpk4" Oct 02 10:26:36 crc kubenswrapper[4934]: I1002 10:26:36.478873 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6gpk4"] Oct 02 10:26:38 crc kubenswrapper[4934]: I1002 10:26:38.381982 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6gpk4" podUID="553f29c5-eb91-4b7e-a5f3-6eddbde93b65" containerName="registry-server" containerID="cri-o://5062ac9255d4cb3d4d61a437e1f801a926b5f03f0f7d94eda107534d7f6b0ce7" gracePeriod=2 Oct 02 10:26:38 crc kubenswrapper[4934]: I1002 10:26:38.861640 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6gpk4" Oct 02 10:26:38 crc kubenswrapper[4934]: I1002 10:26:38.952689 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnmxm\" (UniqueName: \"kubernetes.io/projected/553f29c5-eb91-4b7e-a5f3-6eddbde93b65-kube-api-access-qnmxm\") pod \"553f29c5-eb91-4b7e-a5f3-6eddbde93b65\" (UID: \"553f29c5-eb91-4b7e-a5f3-6eddbde93b65\") " Oct 02 10:26:38 crc kubenswrapper[4934]: I1002 10:26:38.952748 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/553f29c5-eb91-4b7e-a5f3-6eddbde93b65-catalog-content\") pod \"553f29c5-eb91-4b7e-a5f3-6eddbde93b65\" (UID: \"553f29c5-eb91-4b7e-a5f3-6eddbde93b65\") " Oct 02 10:26:38 crc kubenswrapper[4934]: I1002 10:26:38.952779 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/553f29c5-eb91-4b7e-a5f3-6eddbde93b65-utilities\") pod \"553f29c5-eb91-4b7e-a5f3-6eddbde93b65\" (UID: \"553f29c5-eb91-4b7e-a5f3-6eddbde93b65\") " Oct 02 10:26:38 crc kubenswrapper[4934]: I1002 10:26:38.954050 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/553f29c5-eb91-4b7e-a5f3-6eddbde93b65-utilities" (OuterVolumeSpecName: "utilities") pod "553f29c5-eb91-4b7e-a5f3-6eddbde93b65" (UID: "553f29c5-eb91-4b7e-a5f3-6eddbde93b65"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:26:38 crc kubenswrapper[4934]: I1002 10:26:38.965853 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/553f29c5-eb91-4b7e-a5f3-6eddbde93b65-kube-api-access-qnmxm" (OuterVolumeSpecName: "kube-api-access-qnmxm") pod "553f29c5-eb91-4b7e-a5f3-6eddbde93b65" (UID: "553f29c5-eb91-4b7e-a5f3-6eddbde93b65"). InnerVolumeSpecName "kube-api-access-qnmxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:26:39 crc kubenswrapper[4934]: I1002 10:26:39.054455 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnmxm\" (UniqueName: \"kubernetes.io/projected/553f29c5-eb91-4b7e-a5f3-6eddbde93b65-kube-api-access-qnmxm\") on node \"crc\" DevicePath \"\"" Oct 02 10:26:39 crc kubenswrapper[4934]: I1002 10:26:39.054501 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/553f29c5-eb91-4b7e-a5f3-6eddbde93b65-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:26:39 crc kubenswrapper[4934]: I1002 10:26:39.069637 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/553f29c5-eb91-4b7e-a5f3-6eddbde93b65-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "553f29c5-eb91-4b7e-a5f3-6eddbde93b65" (UID: "553f29c5-eb91-4b7e-a5f3-6eddbde93b65"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:26:39 crc kubenswrapper[4934]: I1002 10:26:39.174544 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/553f29c5-eb91-4b7e-a5f3-6eddbde93b65-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:26:39 crc kubenswrapper[4934]: I1002 10:26:39.393649 4934 generic.go:334] "Generic (PLEG): container finished" podID="553f29c5-eb91-4b7e-a5f3-6eddbde93b65" containerID="5062ac9255d4cb3d4d61a437e1f801a926b5f03f0f7d94eda107534d7f6b0ce7" exitCode=0 Oct 02 10:26:39 crc kubenswrapper[4934]: I1002 10:26:39.393712 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6gpk4" event={"ID":"553f29c5-eb91-4b7e-a5f3-6eddbde93b65","Type":"ContainerDied","Data":"5062ac9255d4cb3d4d61a437e1f801a926b5f03f0f7d94eda107534d7f6b0ce7"} Oct 02 10:26:39 crc kubenswrapper[4934]: I1002 10:26:39.393741 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6gpk4" event={"ID":"553f29c5-eb91-4b7e-a5f3-6eddbde93b65","Type":"ContainerDied","Data":"e2dc6d950e2ffc2dc2a39ff83b9f847ca20240d33468d3f47c39984c550c711a"} Oct 02 10:26:39 crc kubenswrapper[4934]: I1002 10:26:39.393757 4934 scope.go:117] "RemoveContainer" containerID="5062ac9255d4cb3d4d61a437e1f801a926b5f03f0f7d94eda107534d7f6b0ce7" Oct 02 10:26:39 crc kubenswrapper[4934]: I1002 10:26:39.393818 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6gpk4" Oct 02 10:26:39 crc kubenswrapper[4934]: I1002 10:26:39.433519 4934 scope.go:117] "RemoveContainer" containerID="b46d13a5f1a0ae5d04ad576a5aa385e70220a85a3f7cd69991b1b1db7cc7148e" Oct 02 10:26:39 crc kubenswrapper[4934]: I1002 10:26:39.464934 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6gpk4"] Oct 02 10:26:39 crc kubenswrapper[4934]: I1002 10:26:39.474877 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6gpk4"] Oct 02 10:26:39 crc kubenswrapper[4934]: I1002 10:26:39.476130 4934 scope.go:117] "RemoveContainer" containerID="f4a1f4c260bd46b6f84dfde602495922395e1640b4ee894a9d68d0c3acf97754" Oct 02 10:26:39 crc kubenswrapper[4934]: I1002 10:26:39.493989 4934 scope.go:117] "RemoveContainer" containerID="5062ac9255d4cb3d4d61a437e1f801a926b5f03f0f7d94eda107534d7f6b0ce7" Oct 02 10:26:39 crc kubenswrapper[4934]: E1002 10:26:39.494484 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5062ac9255d4cb3d4d61a437e1f801a926b5f03f0f7d94eda107534d7f6b0ce7\": container with ID starting with 5062ac9255d4cb3d4d61a437e1f801a926b5f03f0f7d94eda107534d7f6b0ce7 not found: ID does not exist" containerID="5062ac9255d4cb3d4d61a437e1f801a926b5f03f0f7d94eda107534d7f6b0ce7" Oct 02 10:26:39 crc kubenswrapper[4934]: I1002 10:26:39.494515 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5062ac9255d4cb3d4d61a437e1f801a926b5f03f0f7d94eda107534d7f6b0ce7"} err="failed to get container status \"5062ac9255d4cb3d4d61a437e1f801a926b5f03f0f7d94eda107534d7f6b0ce7\": rpc error: code = NotFound desc = could not find container \"5062ac9255d4cb3d4d61a437e1f801a926b5f03f0f7d94eda107534d7f6b0ce7\": container with ID starting with 5062ac9255d4cb3d4d61a437e1f801a926b5f03f0f7d94eda107534d7f6b0ce7 not found: ID does not exist" Oct 02 10:26:39 crc kubenswrapper[4934]: I1002 10:26:39.494535 4934 scope.go:117] "RemoveContainer" containerID="b46d13a5f1a0ae5d04ad576a5aa385e70220a85a3f7cd69991b1b1db7cc7148e" Oct 02 10:26:39 crc kubenswrapper[4934]: E1002 10:26:39.494931 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b46d13a5f1a0ae5d04ad576a5aa385e70220a85a3f7cd69991b1b1db7cc7148e\": container with ID starting with b46d13a5f1a0ae5d04ad576a5aa385e70220a85a3f7cd69991b1b1db7cc7148e not found: ID does not exist" containerID="b46d13a5f1a0ae5d04ad576a5aa385e70220a85a3f7cd69991b1b1db7cc7148e" Oct 02 10:26:39 crc kubenswrapper[4934]: I1002 10:26:39.494981 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b46d13a5f1a0ae5d04ad576a5aa385e70220a85a3f7cd69991b1b1db7cc7148e"} err="failed to get container status \"b46d13a5f1a0ae5d04ad576a5aa385e70220a85a3f7cd69991b1b1db7cc7148e\": rpc error: code = NotFound desc = could not find container \"b46d13a5f1a0ae5d04ad576a5aa385e70220a85a3f7cd69991b1b1db7cc7148e\": container with ID starting with b46d13a5f1a0ae5d04ad576a5aa385e70220a85a3f7cd69991b1b1db7cc7148e not found: ID does not exist" Oct 02 10:26:39 crc kubenswrapper[4934]: I1002 10:26:39.495014 4934 scope.go:117] "RemoveContainer" containerID="f4a1f4c260bd46b6f84dfde602495922395e1640b4ee894a9d68d0c3acf97754" Oct 02 10:26:39 crc kubenswrapper[4934]: E1002 10:26:39.495323 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4a1f4c260bd46b6f84dfde602495922395e1640b4ee894a9d68d0c3acf97754\": container with ID starting with f4a1f4c260bd46b6f84dfde602495922395e1640b4ee894a9d68d0c3acf97754 not found: ID does not exist" containerID="f4a1f4c260bd46b6f84dfde602495922395e1640b4ee894a9d68d0c3acf97754" Oct 02 10:26:39 crc kubenswrapper[4934]: I1002 10:26:39.495360 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4a1f4c260bd46b6f84dfde602495922395e1640b4ee894a9d68d0c3acf97754"} err="failed to get container status \"f4a1f4c260bd46b6f84dfde602495922395e1640b4ee894a9d68d0c3acf97754\": rpc error: code = NotFound desc = could not find container \"f4a1f4c260bd46b6f84dfde602495922395e1640b4ee894a9d68d0c3acf97754\": container with ID starting with f4a1f4c260bd46b6f84dfde602495922395e1640b4ee894a9d68d0c3acf97754 not found: ID does not exist" Oct 02 10:26:40 crc kubenswrapper[4934]: I1002 10:26:40.929114 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="553f29c5-eb91-4b7e-a5f3-6eddbde93b65" path="/var/lib/kubelet/pods/553f29c5-eb91-4b7e-a5f3-6eddbde93b65/volumes" Oct 02 10:26:41 crc kubenswrapper[4934]: I1002 10:26:41.913233 4934 scope.go:117] "RemoveContainer" containerID="7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245" Oct 02 10:26:41 crc kubenswrapper[4934]: E1002 10:26:41.913547 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:26:55 crc kubenswrapper[4934]: I1002 10:26:55.913351 4934 scope.go:117] "RemoveContainer" containerID="7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245" Oct 02 10:26:55 crc kubenswrapper[4934]: E1002 10:26:55.914546 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:27:07 crc kubenswrapper[4934]: I1002 10:27:07.913812 4934 scope.go:117] "RemoveContainer" containerID="7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245" Oct 02 10:27:07 crc kubenswrapper[4934]: E1002 10:27:07.914722 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:27:19 crc kubenswrapper[4934]: I1002 10:27:19.912997 4934 scope.go:117] "RemoveContainer" containerID="7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245" Oct 02 10:27:19 crc kubenswrapper[4934]: E1002 10:27:19.913811 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:27:34 crc kubenswrapper[4934]: I1002 10:27:34.913417 4934 scope.go:117] "RemoveContainer" containerID="7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245" Oct 02 10:27:34 crc kubenswrapper[4934]: E1002 10:27:34.914773 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:27:45 crc kubenswrapper[4934]: I1002 10:27:45.913548 4934 scope.go:117] "RemoveContainer" containerID="7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245" Oct 02 10:27:45 crc kubenswrapper[4934]: E1002 10:27:45.914528 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:27:56 crc kubenswrapper[4934]: I1002 10:27:56.913437 4934 scope.go:117] "RemoveContainer" containerID="7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245" Oct 02 10:27:56 crc kubenswrapper[4934]: E1002 10:27:56.914290 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:28:11 crc kubenswrapper[4934]: I1002 10:28:11.913312 4934 scope.go:117] "RemoveContainer" containerID="7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245" Oct 02 10:28:11 crc kubenswrapper[4934]: E1002 10:28:11.914227 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:28:23 crc kubenswrapper[4934]: I1002 10:28:23.913290 4934 scope.go:117] "RemoveContainer" containerID="7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245" Oct 02 10:28:23 crc kubenswrapper[4934]: E1002 10:28:23.914929 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:28:38 crc kubenswrapper[4934]: I1002 10:28:38.920547 4934 scope.go:117] "RemoveContainer" containerID="7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245" Oct 02 10:28:38 crc kubenswrapper[4934]: E1002 10:28:38.921698 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:28:53 crc kubenswrapper[4934]: I1002 10:28:53.913076 4934 scope.go:117] "RemoveContainer" containerID="7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245" Oct 02 10:28:53 crc kubenswrapper[4934]: E1002 10:28:53.913803 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:29:04 crc kubenswrapper[4934]: I1002 10:29:04.914047 4934 scope.go:117] "RemoveContainer" containerID="7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245" Oct 02 10:29:04 crc kubenswrapper[4934]: E1002 10:29:04.915104 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:29:18 crc kubenswrapper[4934]: I1002 10:29:18.919881 4934 scope.go:117] "RemoveContainer" containerID="7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245" Oct 02 10:29:18 crc kubenswrapper[4934]: E1002 10:29:18.920818 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:29:31 crc kubenswrapper[4934]: I1002 10:29:31.914054 4934 scope.go:117] "RemoveContainer" containerID="7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245" Oct 02 10:29:31 crc kubenswrapper[4934]: E1002 10:29:31.914879 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:29:44 crc kubenswrapper[4934]: I1002 10:29:44.913326 4934 scope.go:117] "RemoveContainer" containerID="7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245" Oct 02 10:29:44 crc kubenswrapper[4934]: E1002 10:29:44.915161 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:29:57 crc kubenswrapper[4934]: I1002 10:29:57.913111 4934 scope.go:117] "RemoveContainer" containerID="7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245" Oct 02 10:29:57 crc kubenswrapper[4934]: E1002 10:29:57.913911 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:30:00 crc kubenswrapper[4934]: I1002 10:30:00.168615 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323350-wh248"] Oct 02 10:30:00 crc kubenswrapper[4934]: E1002 10:30:00.169489 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="553f29c5-eb91-4b7e-a5f3-6eddbde93b65" containerName="extract-utilities" Oct 02 10:30:00 crc kubenswrapper[4934]: I1002 10:30:00.169513 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="553f29c5-eb91-4b7e-a5f3-6eddbde93b65" containerName="extract-utilities" Oct 02 10:30:00 crc kubenswrapper[4934]: E1002 10:30:00.169542 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="553f29c5-eb91-4b7e-a5f3-6eddbde93b65" containerName="extract-content" Oct 02 10:30:00 crc kubenswrapper[4934]: I1002 10:30:00.169553 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="553f29c5-eb91-4b7e-a5f3-6eddbde93b65" containerName="extract-content" Oct 02 10:30:00 crc kubenswrapper[4934]: E1002 10:30:00.169656 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="553f29c5-eb91-4b7e-a5f3-6eddbde93b65" containerName="registry-server" Oct 02 10:30:00 crc kubenswrapper[4934]: I1002 10:30:00.169670 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="553f29c5-eb91-4b7e-a5f3-6eddbde93b65" containerName="registry-server" Oct 02 10:30:00 crc kubenswrapper[4934]: I1002 10:30:00.169919 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="553f29c5-eb91-4b7e-a5f3-6eddbde93b65" containerName="registry-server" Oct 02 10:30:00 crc kubenswrapper[4934]: I1002 10:30:00.170707 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-wh248" Oct 02 10:30:00 crc kubenswrapper[4934]: I1002 10:30:00.174806 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 10:30:00 crc kubenswrapper[4934]: I1002 10:30:00.177967 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323350-wh248"] Oct 02 10:30:00 crc kubenswrapper[4934]: I1002 10:30:00.179725 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 10:30:00 crc kubenswrapper[4934]: I1002 10:30:00.316249 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4a87f0aa-7ba8-44dd-b72d-e0c235640e0d-config-volume\") pod \"collect-profiles-29323350-wh248\" (UID: \"4a87f0aa-7ba8-44dd-b72d-e0c235640e0d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-wh248" Oct 02 10:30:00 crc kubenswrapper[4934]: I1002 10:30:00.316314 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4a87f0aa-7ba8-44dd-b72d-e0c235640e0d-secret-volume\") pod \"collect-profiles-29323350-wh248\" (UID: \"4a87f0aa-7ba8-44dd-b72d-e0c235640e0d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-wh248" Oct 02 10:30:00 crc kubenswrapper[4934]: I1002 10:30:00.316570 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cntsj\" (UniqueName: \"kubernetes.io/projected/4a87f0aa-7ba8-44dd-b72d-e0c235640e0d-kube-api-access-cntsj\") pod \"collect-profiles-29323350-wh248\" (UID: \"4a87f0aa-7ba8-44dd-b72d-e0c235640e0d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-wh248" Oct 02 10:30:00 crc kubenswrapper[4934]: I1002 10:30:00.417641 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4a87f0aa-7ba8-44dd-b72d-e0c235640e0d-config-volume\") pod \"collect-profiles-29323350-wh248\" (UID: \"4a87f0aa-7ba8-44dd-b72d-e0c235640e0d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-wh248" Oct 02 10:30:00 crc kubenswrapper[4934]: I1002 10:30:00.417700 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4a87f0aa-7ba8-44dd-b72d-e0c235640e0d-secret-volume\") pod \"collect-profiles-29323350-wh248\" (UID: \"4a87f0aa-7ba8-44dd-b72d-e0c235640e0d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-wh248" Oct 02 10:30:00 crc kubenswrapper[4934]: I1002 10:30:00.417768 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cntsj\" (UniqueName: \"kubernetes.io/projected/4a87f0aa-7ba8-44dd-b72d-e0c235640e0d-kube-api-access-cntsj\") pod \"collect-profiles-29323350-wh248\" (UID: \"4a87f0aa-7ba8-44dd-b72d-e0c235640e0d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-wh248" Oct 02 10:30:00 crc kubenswrapper[4934]: I1002 10:30:00.419025 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4a87f0aa-7ba8-44dd-b72d-e0c235640e0d-config-volume\") pod \"collect-profiles-29323350-wh248\" (UID: \"4a87f0aa-7ba8-44dd-b72d-e0c235640e0d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-wh248" Oct 02 10:30:00 crc kubenswrapper[4934]: I1002 10:30:00.431234 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4a87f0aa-7ba8-44dd-b72d-e0c235640e0d-secret-volume\") pod \"collect-profiles-29323350-wh248\" (UID: \"4a87f0aa-7ba8-44dd-b72d-e0c235640e0d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-wh248" Oct 02 10:30:00 crc kubenswrapper[4934]: I1002 10:30:00.439028 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cntsj\" (UniqueName: \"kubernetes.io/projected/4a87f0aa-7ba8-44dd-b72d-e0c235640e0d-kube-api-access-cntsj\") pod \"collect-profiles-29323350-wh248\" (UID: \"4a87f0aa-7ba8-44dd-b72d-e0c235640e0d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-wh248" Oct 02 10:30:00 crc kubenswrapper[4934]: I1002 10:30:00.499016 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-wh248" Oct 02 10:30:00 crc kubenswrapper[4934]: I1002 10:30:00.937147 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323350-wh248"] Oct 02 10:30:01 crc kubenswrapper[4934]: I1002 10:30:01.211392 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-wh248" event={"ID":"4a87f0aa-7ba8-44dd-b72d-e0c235640e0d","Type":"ContainerStarted","Data":"53ef02571bceaeaf727a911421e4f6ce17d8e14cc898d3fad39ce0b0dc0ff551"} Oct 02 10:30:01 crc kubenswrapper[4934]: I1002 10:30:01.211432 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-wh248" event={"ID":"4a87f0aa-7ba8-44dd-b72d-e0c235640e0d","Type":"ContainerStarted","Data":"2e3428430b7f8c53ac8c78ea887e97b660487b3003aea9f9109694902f0ef146"} Oct 02 10:30:01 crc kubenswrapper[4934]: I1002 10:30:01.238698 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-wh248" podStartSLOduration=1.238680602 podStartE2EDuration="1.238680602s" podCreationTimestamp="2025-10-02 10:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:30:01.232881411 +0000 UTC m=+2472.985522933" watchObservedRunningTime="2025-10-02 10:30:01.238680602 +0000 UTC m=+2472.991322124" Oct 02 10:30:02 crc kubenswrapper[4934]: I1002 10:30:02.221462 4934 generic.go:334] "Generic (PLEG): container finished" podID="4a87f0aa-7ba8-44dd-b72d-e0c235640e0d" containerID="53ef02571bceaeaf727a911421e4f6ce17d8e14cc898d3fad39ce0b0dc0ff551" exitCode=0 Oct 02 10:30:02 crc kubenswrapper[4934]: I1002 10:30:02.221594 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-wh248" event={"ID":"4a87f0aa-7ba8-44dd-b72d-e0c235640e0d","Type":"ContainerDied","Data":"53ef02571bceaeaf727a911421e4f6ce17d8e14cc898d3fad39ce0b0dc0ff551"} Oct 02 10:30:03 crc kubenswrapper[4934]: I1002 10:30:03.572157 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-wh248" Oct 02 10:30:03 crc kubenswrapper[4934]: I1002 10:30:03.669067 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cntsj\" (UniqueName: \"kubernetes.io/projected/4a87f0aa-7ba8-44dd-b72d-e0c235640e0d-kube-api-access-cntsj\") pod \"4a87f0aa-7ba8-44dd-b72d-e0c235640e0d\" (UID: \"4a87f0aa-7ba8-44dd-b72d-e0c235640e0d\") " Oct 02 10:30:03 crc kubenswrapper[4934]: I1002 10:30:03.669149 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4a87f0aa-7ba8-44dd-b72d-e0c235640e0d-secret-volume\") pod \"4a87f0aa-7ba8-44dd-b72d-e0c235640e0d\" (UID: \"4a87f0aa-7ba8-44dd-b72d-e0c235640e0d\") " Oct 02 10:30:03 crc kubenswrapper[4934]: I1002 10:30:03.669168 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4a87f0aa-7ba8-44dd-b72d-e0c235640e0d-config-volume\") pod \"4a87f0aa-7ba8-44dd-b72d-e0c235640e0d\" (UID: \"4a87f0aa-7ba8-44dd-b72d-e0c235640e0d\") " Oct 02 10:30:03 crc kubenswrapper[4934]: I1002 10:30:03.670019 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a87f0aa-7ba8-44dd-b72d-e0c235640e0d-config-volume" (OuterVolumeSpecName: "config-volume") pod "4a87f0aa-7ba8-44dd-b72d-e0c235640e0d" (UID: "4a87f0aa-7ba8-44dd-b72d-e0c235640e0d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:30:03 crc kubenswrapper[4934]: I1002 10:30:03.674663 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a87f0aa-7ba8-44dd-b72d-e0c235640e0d-kube-api-access-cntsj" (OuterVolumeSpecName: "kube-api-access-cntsj") pod "4a87f0aa-7ba8-44dd-b72d-e0c235640e0d" (UID: "4a87f0aa-7ba8-44dd-b72d-e0c235640e0d"). InnerVolumeSpecName "kube-api-access-cntsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:30:03 crc kubenswrapper[4934]: I1002 10:30:03.674787 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a87f0aa-7ba8-44dd-b72d-e0c235640e0d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4a87f0aa-7ba8-44dd-b72d-e0c235640e0d" (UID: "4a87f0aa-7ba8-44dd-b72d-e0c235640e0d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:30:03 crc kubenswrapper[4934]: I1002 10:30:03.771175 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cntsj\" (UniqueName: \"kubernetes.io/projected/4a87f0aa-7ba8-44dd-b72d-e0c235640e0d-kube-api-access-cntsj\") on node \"crc\" DevicePath \"\"" Oct 02 10:30:03 crc kubenswrapper[4934]: I1002 10:30:03.771232 4934 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4a87f0aa-7ba8-44dd-b72d-e0c235640e0d-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 10:30:03 crc kubenswrapper[4934]: I1002 10:30:03.771251 4934 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4a87f0aa-7ba8-44dd-b72d-e0c235640e0d-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 10:30:04 crc kubenswrapper[4934]: I1002 10:30:04.236247 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-wh248" event={"ID":"4a87f0aa-7ba8-44dd-b72d-e0c235640e0d","Type":"ContainerDied","Data":"2e3428430b7f8c53ac8c78ea887e97b660487b3003aea9f9109694902f0ef146"} Oct 02 10:30:04 crc kubenswrapper[4934]: I1002 10:30:04.236293 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e3428430b7f8c53ac8c78ea887e97b660487b3003aea9f9109694902f0ef146" Oct 02 10:30:04 crc kubenswrapper[4934]: I1002 10:30:04.236339 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323350-wh248" Oct 02 10:30:04 crc kubenswrapper[4934]: I1002 10:30:04.665715 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323305-jv64v"] Oct 02 10:30:04 crc kubenswrapper[4934]: I1002 10:30:04.672344 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323305-jv64v"] Oct 02 10:30:04 crc kubenswrapper[4934]: I1002 10:30:04.925066 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a" path="/var/lib/kubelet/pods/c1daa2ad-e0dd-41aa-87c2-3fd398e00f2a/volumes" Oct 02 10:30:11 crc kubenswrapper[4934]: I1002 10:30:11.914010 4934 scope.go:117] "RemoveContainer" containerID="7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245" Oct 02 10:30:11 crc kubenswrapper[4934]: E1002 10:30:11.914764 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:30:13 crc kubenswrapper[4934]: I1002 10:30:13.311138 4934 scope.go:117] "RemoveContainer" containerID="0de8328c21c16ce1160c7e885fa77ebef1fa96de06d43610aad59115c78bf81e" Oct 02 10:30:26 crc kubenswrapper[4934]: I1002 10:30:26.913854 4934 scope.go:117] "RemoveContainer" containerID="7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245" Oct 02 10:30:26 crc kubenswrapper[4934]: E1002 10:30:26.915168 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:30:38 crc kubenswrapper[4934]: I1002 10:30:38.920315 4934 scope.go:117] "RemoveContainer" containerID="7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245" Oct 02 10:30:39 crc kubenswrapper[4934]: I1002 10:30:39.545560 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerStarted","Data":"4bb852b9106695a7e9f562beaad217ab76737694c80e8ef423f534bbeba47953"} Oct 02 10:31:45 crc kubenswrapper[4934]: I1002 10:31:45.919552 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zw45p"] Oct 02 10:31:45 crc kubenswrapper[4934]: E1002 10:31:45.920376 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a87f0aa-7ba8-44dd-b72d-e0c235640e0d" containerName="collect-profiles" Oct 02 10:31:45 crc kubenswrapper[4934]: I1002 10:31:45.920391 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a87f0aa-7ba8-44dd-b72d-e0c235640e0d" containerName="collect-profiles" Oct 02 10:31:45 crc kubenswrapper[4934]: I1002 10:31:45.920551 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a87f0aa-7ba8-44dd-b72d-e0c235640e0d" containerName="collect-profiles" Oct 02 10:31:45 crc kubenswrapper[4934]: I1002 10:31:45.921569 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zw45p" Oct 02 10:31:45 crc kubenswrapper[4934]: I1002 10:31:45.935490 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zw45p"] Oct 02 10:31:45 crc kubenswrapper[4934]: I1002 10:31:45.998283 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b898bbb6-5629-469e-81e1-2d56e85b0415-utilities\") pod \"community-operators-zw45p\" (UID: \"b898bbb6-5629-469e-81e1-2d56e85b0415\") " pod="openshift-marketplace/community-operators-zw45p" Oct 02 10:31:45 crc kubenswrapper[4934]: I1002 10:31:45.998378 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcrsd\" (UniqueName: \"kubernetes.io/projected/b898bbb6-5629-469e-81e1-2d56e85b0415-kube-api-access-gcrsd\") pod \"community-operators-zw45p\" (UID: \"b898bbb6-5629-469e-81e1-2d56e85b0415\") " pod="openshift-marketplace/community-operators-zw45p" Oct 02 10:31:45 crc kubenswrapper[4934]: I1002 10:31:45.998476 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b898bbb6-5629-469e-81e1-2d56e85b0415-catalog-content\") pod \"community-operators-zw45p\" (UID: \"b898bbb6-5629-469e-81e1-2d56e85b0415\") " pod="openshift-marketplace/community-operators-zw45p" Oct 02 10:31:46 crc kubenswrapper[4934]: I1002 10:31:46.100333 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b898bbb6-5629-469e-81e1-2d56e85b0415-utilities\") pod \"community-operators-zw45p\" (UID: \"b898bbb6-5629-469e-81e1-2d56e85b0415\") " pod="openshift-marketplace/community-operators-zw45p" Oct 02 10:31:46 crc kubenswrapper[4934]: I1002 10:31:46.100423 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcrsd\" (UniqueName: \"kubernetes.io/projected/b898bbb6-5629-469e-81e1-2d56e85b0415-kube-api-access-gcrsd\") pod \"community-operators-zw45p\" (UID: \"b898bbb6-5629-469e-81e1-2d56e85b0415\") " pod="openshift-marketplace/community-operators-zw45p" Oct 02 10:31:46 crc kubenswrapper[4934]: I1002 10:31:46.100471 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b898bbb6-5629-469e-81e1-2d56e85b0415-catalog-content\") pod \"community-operators-zw45p\" (UID: \"b898bbb6-5629-469e-81e1-2d56e85b0415\") " pod="openshift-marketplace/community-operators-zw45p" Oct 02 10:31:46 crc kubenswrapper[4934]: I1002 10:31:46.101034 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b898bbb6-5629-469e-81e1-2d56e85b0415-catalog-content\") pod \"community-operators-zw45p\" (UID: \"b898bbb6-5629-469e-81e1-2d56e85b0415\") " pod="openshift-marketplace/community-operators-zw45p" Oct 02 10:31:46 crc kubenswrapper[4934]: I1002 10:31:46.101294 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b898bbb6-5629-469e-81e1-2d56e85b0415-utilities\") pod \"community-operators-zw45p\" (UID: \"b898bbb6-5629-469e-81e1-2d56e85b0415\") " pod="openshift-marketplace/community-operators-zw45p" Oct 02 10:31:46 crc kubenswrapper[4934]: I1002 10:31:46.123273 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcrsd\" (UniqueName: \"kubernetes.io/projected/b898bbb6-5629-469e-81e1-2d56e85b0415-kube-api-access-gcrsd\") pod \"community-operators-zw45p\" (UID: \"b898bbb6-5629-469e-81e1-2d56e85b0415\") " pod="openshift-marketplace/community-operators-zw45p" Oct 02 10:31:46 crc kubenswrapper[4934]: I1002 10:31:46.253778 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zw45p" Oct 02 10:31:46 crc kubenswrapper[4934]: I1002 10:31:46.743922 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zw45p"] Oct 02 10:31:47 crc kubenswrapper[4934]: I1002 10:31:47.094739 4934 generic.go:334] "Generic (PLEG): container finished" podID="b898bbb6-5629-469e-81e1-2d56e85b0415" containerID="fa41d9f59be324374554bc4396f55f3e8cf456dd8140718eb51db3bf42b08981" exitCode=0 Oct 02 10:31:47 crc kubenswrapper[4934]: I1002 10:31:47.094798 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zw45p" event={"ID":"b898bbb6-5629-469e-81e1-2d56e85b0415","Type":"ContainerDied","Data":"fa41d9f59be324374554bc4396f55f3e8cf456dd8140718eb51db3bf42b08981"} Oct 02 10:31:47 crc kubenswrapper[4934]: I1002 10:31:47.095014 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zw45p" event={"ID":"b898bbb6-5629-469e-81e1-2d56e85b0415","Type":"ContainerStarted","Data":"1b9a57e1da46c492381a61045b0825990aca79290c45bbf0717663e1f2e635c2"} Oct 02 10:31:47 crc kubenswrapper[4934]: I1002 10:31:47.096556 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 10:31:48 crc kubenswrapper[4934]: I1002 10:31:48.114036 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zw45p" event={"ID":"b898bbb6-5629-469e-81e1-2d56e85b0415","Type":"ContainerStarted","Data":"0be362d04af8f14aca4462593be76685b80972bd1d3b5d1bf58e7dfe0929ba9c"} Oct 02 10:31:49 crc kubenswrapper[4934]: I1002 10:31:49.125939 4934 generic.go:334] "Generic (PLEG): container finished" podID="b898bbb6-5629-469e-81e1-2d56e85b0415" containerID="0be362d04af8f14aca4462593be76685b80972bd1d3b5d1bf58e7dfe0929ba9c" exitCode=0 Oct 02 10:31:49 crc kubenswrapper[4934]: I1002 10:31:49.125991 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zw45p" event={"ID":"b898bbb6-5629-469e-81e1-2d56e85b0415","Type":"ContainerDied","Data":"0be362d04af8f14aca4462593be76685b80972bd1d3b5d1bf58e7dfe0929ba9c"} Oct 02 10:31:50 crc kubenswrapper[4934]: I1002 10:31:50.137088 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zw45p" event={"ID":"b898bbb6-5629-469e-81e1-2d56e85b0415","Type":"ContainerStarted","Data":"0097dcfcb44a0256aec9338cbc763782189ce18506cc6bb22b4d8e16b8624d2e"} Oct 02 10:31:50 crc kubenswrapper[4934]: I1002 10:31:50.164163 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zw45p" podStartSLOduration=2.572492934 podStartE2EDuration="5.164138504s" podCreationTimestamp="2025-10-02 10:31:45 +0000 UTC" firstStartedPulling="2025-10-02 10:31:47.096103162 +0000 UTC m=+2578.848744724" lastFinishedPulling="2025-10-02 10:31:49.687748772 +0000 UTC m=+2581.440390294" observedRunningTime="2025-10-02 10:31:50.159548208 +0000 UTC m=+2581.912189770" watchObservedRunningTime="2025-10-02 10:31:50.164138504 +0000 UTC m=+2581.916780036" Oct 02 10:31:56 crc kubenswrapper[4934]: I1002 10:31:56.254979 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zw45p" Oct 02 10:31:56 crc kubenswrapper[4934]: I1002 10:31:56.255753 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zw45p" Oct 02 10:31:56 crc kubenswrapper[4934]: I1002 10:31:56.327348 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zw45p" Oct 02 10:31:57 crc kubenswrapper[4934]: I1002 10:31:57.261847 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zw45p" Oct 02 10:31:57 crc kubenswrapper[4934]: I1002 10:31:57.327745 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zw45p"] Oct 02 10:31:59 crc kubenswrapper[4934]: I1002 10:31:59.205906 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zw45p" podUID="b898bbb6-5629-469e-81e1-2d56e85b0415" containerName="registry-server" containerID="cri-o://0097dcfcb44a0256aec9338cbc763782189ce18506cc6bb22b4d8e16b8624d2e" gracePeriod=2 Oct 02 10:31:59 crc kubenswrapper[4934]: I1002 10:31:59.585569 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zw45p" Oct 02 10:31:59 crc kubenswrapper[4934]: I1002 10:31:59.699139 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b898bbb6-5629-469e-81e1-2d56e85b0415-utilities\") pod \"b898bbb6-5629-469e-81e1-2d56e85b0415\" (UID: \"b898bbb6-5629-469e-81e1-2d56e85b0415\") " Oct 02 10:31:59 crc kubenswrapper[4934]: I1002 10:31:59.699362 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcrsd\" (UniqueName: \"kubernetes.io/projected/b898bbb6-5629-469e-81e1-2d56e85b0415-kube-api-access-gcrsd\") pod \"b898bbb6-5629-469e-81e1-2d56e85b0415\" (UID: \"b898bbb6-5629-469e-81e1-2d56e85b0415\") " Oct 02 10:31:59 crc kubenswrapper[4934]: I1002 10:31:59.699431 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b898bbb6-5629-469e-81e1-2d56e85b0415-catalog-content\") pod \"b898bbb6-5629-469e-81e1-2d56e85b0415\" (UID: \"b898bbb6-5629-469e-81e1-2d56e85b0415\") " Oct 02 10:31:59 crc kubenswrapper[4934]: I1002 10:31:59.700302 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b898bbb6-5629-469e-81e1-2d56e85b0415-utilities" (OuterVolumeSpecName: "utilities") pod "b898bbb6-5629-469e-81e1-2d56e85b0415" (UID: "b898bbb6-5629-469e-81e1-2d56e85b0415"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:31:59 crc kubenswrapper[4934]: I1002 10:31:59.704302 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b898bbb6-5629-469e-81e1-2d56e85b0415-kube-api-access-gcrsd" (OuterVolumeSpecName: "kube-api-access-gcrsd") pod "b898bbb6-5629-469e-81e1-2d56e85b0415" (UID: "b898bbb6-5629-469e-81e1-2d56e85b0415"). InnerVolumeSpecName "kube-api-access-gcrsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:31:59 crc kubenswrapper[4934]: I1002 10:31:59.755745 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b898bbb6-5629-469e-81e1-2d56e85b0415-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b898bbb6-5629-469e-81e1-2d56e85b0415" (UID: "b898bbb6-5629-469e-81e1-2d56e85b0415"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:31:59 crc kubenswrapper[4934]: I1002 10:31:59.801558 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcrsd\" (UniqueName: \"kubernetes.io/projected/b898bbb6-5629-469e-81e1-2d56e85b0415-kube-api-access-gcrsd\") on node \"crc\" DevicePath \"\"" Oct 02 10:31:59 crc kubenswrapper[4934]: I1002 10:31:59.801633 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b898bbb6-5629-469e-81e1-2d56e85b0415-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:31:59 crc kubenswrapper[4934]: I1002 10:31:59.801651 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b898bbb6-5629-469e-81e1-2d56e85b0415-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:32:00 crc kubenswrapper[4934]: I1002 10:32:00.214940 4934 generic.go:334] "Generic (PLEG): container finished" podID="b898bbb6-5629-469e-81e1-2d56e85b0415" containerID="0097dcfcb44a0256aec9338cbc763782189ce18506cc6bb22b4d8e16b8624d2e" exitCode=0 Oct 02 10:32:00 crc kubenswrapper[4934]: I1002 10:32:00.215146 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zw45p" event={"ID":"b898bbb6-5629-469e-81e1-2d56e85b0415","Type":"ContainerDied","Data":"0097dcfcb44a0256aec9338cbc763782189ce18506cc6bb22b4d8e16b8624d2e"} Oct 02 10:32:00 crc kubenswrapper[4934]: I1002 10:32:00.215252 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zw45p" Oct 02 10:32:00 crc kubenswrapper[4934]: I1002 10:32:00.215295 4934 scope.go:117] "RemoveContainer" containerID="0097dcfcb44a0256aec9338cbc763782189ce18506cc6bb22b4d8e16b8624d2e" Oct 02 10:32:00 crc kubenswrapper[4934]: I1002 10:32:00.215271 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zw45p" event={"ID":"b898bbb6-5629-469e-81e1-2d56e85b0415","Type":"ContainerDied","Data":"1b9a57e1da46c492381a61045b0825990aca79290c45bbf0717663e1f2e635c2"} Oct 02 10:32:00 crc kubenswrapper[4934]: I1002 10:32:00.251031 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zw45p"] Oct 02 10:32:00 crc kubenswrapper[4934]: I1002 10:32:00.257171 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zw45p"] Oct 02 10:32:00 crc kubenswrapper[4934]: I1002 10:32:00.260116 4934 scope.go:117] "RemoveContainer" containerID="0be362d04af8f14aca4462593be76685b80972bd1d3b5d1bf58e7dfe0929ba9c" Oct 02 10:32:00 crc kubenswrapper[4934]: I1002 10:32:00.280954 4934 scope.go:117] "RemoveContainer" containerID="fa41d9f59be324374554bc4396f55f3e8cf456dd8140718eb51db3bf42b08981" Oct 02 10:32:00 crc kubenswrapper[4934]: I1002 10:32:00.321729 4934 scope.go:117] "RemoveContainer" containerID="0097dcfcb44a0256aec9338cbc763782189ce18506cc6bb22b4d8e16b8624d2e" Oct 02 10:32:00 crc kubenswrapper[4934]: E1002 10:32:00.322020 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0097dcfcb44a0256aec9338cbc763782189ce18506cc6bb22b4d8e16b8624d2e\": container with ID starting with 0097dcfcb44a0256aec9338cbc763782189ce18506cc6bb22b4d8e16b8624d2e not found: ID does not exist" containerID="0097dcfcb44a0256aec9338cbc763782189ce18506cc6bb22b4d8e16b8624d2e" Oct 02 10:32:00 crc kubenswrapper[4934]: I1002 10:32:00.322061 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0097dcfcb44a0256aec9338cbc763782189ce18506cc6bb22b4d8e16b8624d2e"} err="failed to get container status \"0097dcfcb44a0256aec9338cbc763782189ce18506cc6bb22b4d8e16b8624d2e\": rpc error: code = NotFound desc = could not find container \"0097dcfcb44a0256aec9338cbc763782189ce18506cc6bb22b4d8e16b8624d2e\": container with ID starting with 0097dcfcb44a0256aec9338cbc763782189ce18506cc6bb22b4d8e16b8624d2e not found: ID does not exist" Oct 02 10:32:00 crc kubenswrapper[4934]: I1002 10:32:00.322087 4934 scope.go:117] "RemoveContainer" containerID="0be362d04af8f14aca4462593be76685b80972bd1d3b5d1bf58e7dfe0929ba9c" Oct 02 10:32:00 crc kubenswrapper[4934]: E1002 10:32:00.322289 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0be362d04af8f14aca4462593be76685b80972bd1d3b5d1bf58e7dfe0929ba9c\": container with ID starting with 0be362d04af8f14aca4462593be76685b80972bd1d3b5d1bf58e7dfe0929ba9c not found: ID does not exist" containerID="0be362d04af8f14aca4462593be76685b80972bd1d3b5d1bf58e7dfe0929ba9c" Oct 02 10:32:00 crc kubenswrapper[4934]: I1002 10:32:00.322315 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0be362d04af8f14aca4462593be76685b80972bd1d3b5d1bf58e7dfe0929ba9c"} err="failed to get container status \"0be362d04af8f14aca4462593be76685b80972bd1d3b5d1bf58e7dfe0929ba9c\": rpc error: code = NotFound desc = could not find container \"0be362d04af8f14aca4462593be76685b80972bd1d3b5d1bf58e7dfe0929ba9c\": container with ID starting with 0be362d04af8f14aca4462593be76685b80972bd1d3b5d1bf58e7dfe0929ba9c not found: ID does not exist" Oct 02 10:32:00 crc kubenswrapper[4934]: I1002 10:32:00.322333 4934 scope.go:117] "RemoveContainer" containerID="fa41d9f59be324374554bc4396f55f3e8cf456dd8140718eb51db3bf42b08981" Oct 02 10:32:00 crc kubenswrapper[4934]: E1002 10:32:00.322511 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa41d9f59be324374554bc4396f55f3e8cf456dd8140718eb51db3bf42b08981\": container with ID starting with fa41d9f59be324374554bc4396f55f3e8cf456dd8140718eb51db3bf42b08981 not found: ID does not exist" containerID="fa41d9f59be324374554bc4396f55f3e8cf456dd8140718eb51db3bf42b08981" Oct 02 10:32:00 crc kubenswrapper[4934]: I1002 10:32:00.322532 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa41d9f59be324374554bc4396f55f3e8cf456dd8140718eb51db3bf42b08981"} err="failed to get container status \"fa41d9f59be324374554bc4396f55f3e8cf456dd8140718eb51db3bf42b08981\": rpc error: code = NotFound desc = could not find container \"fa41d9f59be324374554bc4396f55f3e8cf456dd8140718eb51db3bf42b08981\": container with ID starting with fa41d9f59be324374554bc4396f55f3e8cf456dd8140718eb51db3bf42b08981 not found: ID does not exist" Oct 02 10:32:00 crc kubenswrapper[4934]: I1002 10:32:00.945190 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b898bbb6-5629-469e-81e1-2d56e85b0415" path="/var/lib/kubelet/pods/b898bbb6-5629-469e-81e1-2d56e85b0415/volumes" Oct 02 10:33:08 crc kubenswrapper[4934]: I1002 10:33:08.439704 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:33:08 crc kubenswrapper[4934]: I1002 10:33:08.442220 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:33:38 crc kubenswrapper[4934]: I1002 10:33:38.439241 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:33:38 crc kubenswrapper[4934]: I1002 10:33:38.440106 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:34:08 crc kubenswrapper[4934]: I1002 10:34:08.440291 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:34:08 crc kubenswrapper[4934]: I1002 10:34:08.441044 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:34:08 crc kubenswrapper[4934]: I1002 10:34:08.441127 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 10:34:08 crc kubenswrapper[4934]: I1002 10:34:08.442087 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4bb852b9106695a7e9f562beaad217ab76737694c80e8ef423f534bbeba47953"} pod="openshift-machine-config-operator/machine-config-daemon-djh5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 10:34:08 crc kubenswrapper[4934]: I1002 10:34:08.442185 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" containerID="cri-o://4bb852b9106695a7e9f562beaad217ab76737694c80e8ef423f534bbeba47953" gracePeriod=600 Oct 02 10:34:09 crc kubenswrapper[4934]: I1002 10:34:09.369172 4934 generic.go:334] "Generic (PLEG): container finished" podID="71db06ef-05b0-4f58-b251-b27117a8500a" containerID="4bb852b9106695a7e9f562beaad217ab76737694c80e8ef423f534bbeba47953" exitCode=0 Oct 02 10:34:09 crc kubenswrapper[4934]: I1002 10:34:09.369245 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerDied","Data":"4bb852b9106695a7e9f562beaad217ab76737694c80e8ef423f534bbeba47953"} Oct 02 10:34:09 crc kubenswrapper[4934]: I1002 10:34:09.369746 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerStarted","Data":"4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5"} Oct 02 10:34:09 crc kubenswrapper[4934]: I1002 10:34:09.369787 4934 scope.go:117] "RemoveContainer" containerID="7525339ef4d6342d8ca2dd3e9e9f80f900e2b98c265f27f77c834081a806b245" Oct 02 10:35:15 crc kubenswrapper[4934]: I1002 10:35:15.407813 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v6p8z"] Oct 02 10:35:15 crc kubenswrapper[4934]: E1002 10:35:15.410493 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b898bbb6-5629-469e-81e1-2d56e85b0415" containerName="registry-server" Oct 02 10:35:15 crc kubenswrapper[4934]: I1002 10:35:15.410527 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b898bbb6-5629-469e-81e1-2d56e85b0415" containerName="registry-server" Oct 02 10:35:15 crc kubenswrapper[4934]: E1002 10:35:15.410547 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b898bbb6-5629-469e-81e1-2d56e85b0415" containerName="extract-utilities" Oct 02 10:35:15 crc kubenswrapper[4934]: I1002 10:35:15.410556 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b898bbb6-5629-469e-81e1-2d56e85b0415" containerName="extract-utilities" Oct 02 10:35:15 crc kubenswrapper[4934]: E1002 10:35:15.410591 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b898bbb6-5629-469e-81e1-2d56e85b0415" containerName="extract-content" Oct 02 10:35:15 crc kubenswrapper[4934]: I1002 10:35:15.410601 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b898bbb6-5629-469e-81e1-2d56e85b0415" containerName="extract-content" Oct 02 10:35:15 crc kubenswrapper[4934]: I1002 10:35:15.410779 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="b898bbb6-5629-469e-81e1-2d56e85b0415" containerName="registry-server" Oct 02 10:35:15 crc kubenswrapper[4934]: I1002 10:35:15.412098 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v6p8z" Oct 02 10:35:15 crc kubenswrapper[4934]: I1002 10:35:15.415062 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v6p8z"] Oct 02 10:35:15 crc kubenswrapper[4934]: I1002 10:35:15.542435 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2642e71-e57e-49df-ae69-318c2e6b06ec-utilities\") pod \"certified-operators-v6p8z\" (UID: \"b2642e71-e57e-49df-ae69-318c2e6b06ec\") " pod="openshift-marketplace/certified-operators-v6p8z" Oct 02 10:35:15 crc kubenswrapper[4934]: I1002 10:35:15.542506 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2642e71-e57e-49df-ae69-318c2e6b06ec-catalog-content\") pod \"certified-operators-v6p8z\" (UID: \"b2642e71-e57e-49df-ae69-318c2e6b06ec\") " pod="openshift-marketplace/certified-operators-v6p8z" Oct 02 10:35:15 crc kubenswrapper[4934]: I1002 10:35:15.542548 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzc4m\" (UniqueName: \"kubernetes.io/projected/b2642e71-e57e-49df-ae69-318c2e6b06ec-kube-api-access-bzc4m\") pod \"certified-operators-v6p8z\" (UID: \"b2642e71-e57e-49df-ae69-318c2e6b06ec\") " pod="openshift-marketplace/certified-operators-v6p8z" Oct 02 10:35:15 crc kubenswrapper[4934]: I1002 10:35:15.644213 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2642e71-e57e-49df-ae69-318c2e6b06ec-utilities\") pod \"certified-operators-v6p8z\" (UID: \"b2642e71-e57e-49df-ae69-318c2e6b06ec\") " pod="openshift-marketplace/certified-operators-v6p8z" Oct 02 10:35:15 crc kubenswrapper[4934]: I1002 10:35:15.644298 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2642e71-e57e-49df-ae69-318c2e6b06ec-catalog-content\") pod \"certified-operators-v6p8z\" (UID: \"b2642e71-e57e-49df-ae69-318c2e6b06ec\") " pod="openshift-marketplace/certified-operators-v6p8z" Oct 02 10:35:15 crc kubenswrapper[4934]: I1002 10:35:15.644356 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzc4m\" (UniqueName: \"kubernetes.io/projected/b2642e71-e57e-49df-ae69-318c2e6b06ec-kube-api-access-bzc4m\") pod \"certified-operators-v6p8z\" (UID: \"b2642e71-e57e-49df-ae69-318c2e6b06ec\") " pod="openshift-marketplace/certified-operators-v6p8z" Oct 02 10:35:15 crc kubenswrapper[4934]: I1002 10:35:15.644927 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2642e71-e57e-49df-ae69-318c2e6b06ec-utilities\") pod \"certified-operators-v6p8z\" (UID: \"b2642e71-e57e-49df-ae69-318c2e6b06ec\") " pod="openshift-marketplace/certified-operators-v6p8z" Oct 02 10:35:15 crc kubenswrapper[4934]: I1002 10:35:15.644999 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2642e71-e57e-49df-ae69-318c2e6b06ec-catalog-content\") pod \"certified-operators-v6p8z\" (UID: \"b2642e71-e57e-49df-ae69-318c2e6b06ec\") " pod="openshift-marketplace/certified-operators-v6p8z" Oct 02 10:35:15 crc kubenswrapper[4934]: I1002 10:35:15.672158 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzc4m\" (UniqueName: \"kubernetes.io/projected/b2642e71-e57e-49df-ae69-318c2e6b06ec-kube-api-access-bzc4m\") pod \"certified-operators-v6p8z\" (UID: \"b2642e71-e57e-49df-ae69-318c2e6b06ec\") " pod="openshift-marketplace/certified-operators-v6p8z" Oct 02 10:35:15 crc kubenswrapper[4934]: I1002 10:35:15.730779 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v6p8z" Oct 02 10:35:16 crc kubenswrapper[4934]: I1002 10:35:16.210775 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v6p8z"] Oct 02 10:35:16 crc kubenswrapper[4934]: I1002 10:35:16.930070 4934 generic.go:334] "Generic (PLEG): container finished" podID="b2642e71-e57e-49df-ae69-318c2e6b06ec" containerID="193263e3ad2d77b5b27bd97e0904c40723d8b757ab586e11e6c9a4c118c04f9b" exitCode=0 Oct 02 10:35:16 crc kubenswrapper[4934]: I1002 10:35:16.930149 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v6p8z" event={"ID":"b2642e71-e57e-49df-ae69-318c2e6b06ec","Type":"ContainerDied","Data":"193263e3ad2d77b5b27bd97e0904c40723d8b757ab586e11e6c9a4c118c04f9b"} Oct 02 10:35:16 crc kubenswrapper[4934]: I1002 10:35:16.930388 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v6p8z" event={"ID":"b2642e71-e57e-49df-ae69-318c2e6b06ec","Type":"ContainerStarted","Data":"be8a6f3dc1c9b0b0e85411c6c16bb2cb2671b1e83d9108b85937a853c9c04a6a"} Oct 02 10:35:18 crc kubenswrapper[4934]: I1002 10:35:18.947951 4934 generic.go:334] "Generic (PLEG): container finished" podID="b2642e71-e57e-49df-ae69-318c2e6b06ec" containerID="76980a305a2fe54a3f41cd58814d5863a7f45964f63edaa991546038aa7c993d" exitCode=0 Oct 02 10:35:18 crc kubenswrapper[4934]: I1002 10:35:18.947988 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v6p8z" event={"ID":"b2642e71-e57e-49df-ae69-318c2e6b06ec","Type":"ContainerDied","Data":"76980a305a2fe54a3f41cd58814d5863a7f45964f63edaa991546038aa7c993d"} Oct 02 10:35:19 crc kubenswrapper[4934]: I1002 10:35:19.959035 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v6p8z" event={"ID":"b2642e71-e57e-49df-ae69-318c2e6b06ec","Type":"ContainerStarted","Data":"327fb065e85f7b8f30afc532378584eca7ecba7ea0ba07312c5099f7adf3fea3"} Oct 02 10:35:19 crc kubenswrapper[4934]: I1002 10:35:19.984321 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v6p8z" podStartSLOduration=2.333275405 podStartE2EDuration="4.984292797s" podCreationTimestamp="2025-10-02 10:35:15 +0000 UTC" firstStartedPulling="2025-10-02 10:35:16.931980209 +0000 UTC m=+2788.684621731" lastFinishedPulling="2025-10-02 10:35:19.582997601 +0000 UTC m=+2791.335639123" observedRunningTime="2025-10-02 10:35:19.975273038 +0000 UTC m=+2791.727914570" watchObservedRunningTime="2025-10-02 10:35:19.984292797 +0000 UTC m=+2791.736934349" Oct 02 10:35:25 crc kubenswrapper[4934]: I1002 10:35:25.731882 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v6p8z" Oct 02 10:35:25 crc kubenswrapper[4934]: I1002 10:35:25.732221 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v6p8z" Oct 02 10:35:25 crc kubenswrapper[4934]: I1002 10:35:25.794242 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v6p8z" Oct 02 10:35:26 crc kubenswrapper[4934]: I1002 10:35:26.063642 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v6p8z" Oct 02 10:35:26 crc kubenswrapper[4934]: I1002 10:35:26.122814 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v6p8z"] Oct 02 10:35:28 crc kubenswrapper[4934]: I1002 10:35:28.021144 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-v6p8z" podUID="b2642e71-e57e-49df-ae69-318c2e6b06ec" containerName="registry-server" containerID="cri-o://327fb065e85f7b8f30afc532378584eca7ecba7ea0ba07312c5099f7adf3fea3" gracePeriod=2 Oct 02 10:35:29 crc kubenswrapper[4934]: I1002 10:35:29.032839 4934 generic.go:334] "Generic (PLEG): container finished" podID="b2642e71-e57e-49df-ae69-318c2e6b06ec" containerID="327fb065e85f7b8f30afc532378584eca7ecba7ea0ba07312c5099f7adf3fea3" exitCode=0 Oct 02 10:35:29 crc kubenswrapper[4934]: I1002 10:35:29.032969 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v6p8z" event={"ID":"b2642e71-e57e-49df-ae69-318c2e6b06ec","Type":"ContainerDied","Data":"327fb065e85f7b8f30afc532378584eca7ecba7ea0ba07312c5099f7adf3fea3"} Oct 02 10:35:29 crc kubenswrapper[4934]: I1002 10:35:29.573670 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v6p8z" Oct 02 10:35:29 crc kubenswrapper[4934]: I1002 10:35:29.668352 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2642e71-e57e-49df-ae69-318c2e6b06ec-utilities\") pod \"b2642e71-e57e-49df-ae69-318c2e6b06ec\" (UID: \"b2642e71-e57e-49df-ae69-318c2e6b06ec\") " Oct 02 10:35:29 crc kubenswrapper[4934]: I1002 10:35:29.668427 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2642e71-e57e-49df-ae69-318c2e6b06ec-catalog-content\") pod \"b2642e71-e57e-49df-ae69-318c2e6b06ec\" (UID: \"b2642e71-e57e-49df-ae69-318c2e6b06ec\") " Oct 02 10:35:29 crc kubenswrapper[4934]: I1002 10:35:29.668474 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzc4m\" (UniqueName: \"kubernetes.io/projected/b2642e71-e57e-49df-ae69-318c2e6b06ec-kube-api-access-bzc4m\") pod \"b2642e71-e57e-49df-ae69-318c2e6b06ec\" (UID: \"b2642e71-e57e-49df-ae69-318c2e6b06ec\") " Oct 02 10:35:29 crc kubenswrapper[4934]: I1002 10:35:29.669260 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2642e71-e57e-49df-ae69-318c2e6b06ec-utilities" (OuterVolumeSpecName: "utilities") pod "b2642e71-e57e-49df-ae69-318c2e6b06ec" (UID: "b2642e71-e57e-49df-ae69-318c2e6b06ec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:35:29 crc kubenswrapper[4934]: I1002 10:35:29.677069 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2642e71-e57e-49df-ae69-318c2e6b06ec-kube-api-access-bzc4m" (OuterVolumeSpecName: "kube-api-access-bzc4m") pod "b2642e71-e57e-49df-ae69-318c2e6b06ec" (UID: "b2642e71-e57e-49df-ae69-318c2e6b06ec"). InnerVolumeSpecName "kube-api-access-bzc4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:35:29 crc kubenswrapper[4934]: I1002 10:35:29.709733 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2642e71-e57e-49df-ae69-318c2e6b06ec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b2642e71-e57e-49df-ae69-318c2e6b06ec" (UID: "b2642e71-e57e-49df-ae69-318c2e6b06ec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:35:29 crc kubenswrapper[4934]: I1002 10:35:29.770493 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2642e71-e57e-49df-ae69-318c2e6b06ec-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:35:29 crc kubenswrapper[4934]: I1002 10:35:29.770755 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2642e71-e57e-49df-ae69-318c2e6b06ec-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:35:29 crc kubenswrapper[4934]: I1002 10:35:29.770851 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzc4m\" (UniqueName: \"kubernetes.io/projected/b2642e71-e57e-49df-ae69-318c2e6b06ec-kube-api-access-bzc4m\") on node \"crc\" DevicePath \"\"" Oct 02 10:35:30 crc kubenswrapper[4934]: I1002 10:35:30.041688 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v6p8z" event={"ID":"b2642e71-e57e-49df-ae69-318c2e6b06ec","Type":"ContainerDied","Data":"be8a6f3dc1c9b0b0e85411c6c16bb2cb2671b1e83d9108b85937a853c9c04a6a"} Oct 02 10:35:30 crc kubenswrapper[4934]: I1002 10:35:30.041734 4934 scope.go:117] "RemoveContainer" containerID="327fb065e85f7b8f30afc532378584eca7ecba7ea0ba07312c5099f7adf3fea3" Oct 02 10:35:30 crc kubenswrapper[4934]: I1002 10:35:30.041746 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v6p8z" Oct 02 10:35:30 crc kubenswrapper[4934]: I1002 10:35:30.080755 4934 scope.go:117] "RemoveContainer" containerID="76980a305a2fe54a3f41cd58814d5863a7f45964f63edaa991546038aa7c993d" Oct 02 10:35:30 crc kubenswrapper[4934]: I1002 10:35:30.084980 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v6p8z"] Oct 02 10:35:30 crc kubenswrapper[4934]: I1002 10:35:30.089288 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-v6p8z"] Oct 02 10:35:30 crc kubenswrapper[4934]: I1002 10:35:30.125471 4934 scope.go:117] "RemoveContainer" containerID="193263e3ad2d77b5b27bd97e0904c40723d8b757ab586e11e6c9a4c118c04f9b" Oct 02 10:35:30 crc kubenswrapper[4934]: I1002 10:35:30.924177 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2642e71-e57e-49df-ae69-318c2e6b06ec" path="/var/lib/kubelet/pods/b2642e71-e57e-49df-ae69-318c2e6b06ec/volumes" Oct 02 10:36:08 crc kubenswrapper[4934]: I1002 10:36:08.440164 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:36:08 crc kubenswrapper[4934]: I1002 10:36:08.440983 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:36:38 crc kubenswrapper[4934]: I1002 10:36:38.439544 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:36:38 crc kubenswrapper[4934]: I1002 10:36:38.440170 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:36:57 crc kubenswrapper[4934]: I1002 10:36:57.135771 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-swl7r"] Oct 02 10:36:57 crc kubenswrapper[4934]: E1002 10:36:57.136501 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2642e71-e57e-49df-ae69-318c2e6b06ec" containerName="extract-utilities" Oct 02 10:36:57 crc kubenswrapper[4934]: I1002 10:36:57.136512 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2642e71-e57e-49df-ae69-318c2e6b06ec" containerName="extract-utilities" Oct 02 10:36:57 crc kubenswrapper[4934]: E1002 10:36:57.136534 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2642e71-e57e-49df-ae69-318c2e6b06ec" containerName="registry-server" Oct 02 10:36:57 crc kubenswrapper[4934]: I1002 10:36:57.136541 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2642e71-e57e-49df-ae69-318c2e6b06ec" containerName="registry-server" Oct 02 10:36:57 crc kubenswrapper[4934]: E1002 10:36:57.136550 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2642e71-e57e-49df-ae69-318c2e6b06ec" containerName="extract-content" Oct 02 10:36:57 crc kubenswrapper[4934]: I1002 10:36:57.136561 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2642e71-e57e-49df-ae69-318c2e6b06ec" containerName="extract-content" Oct 02 10:36:57 crc kubenswrapper[4934]: I1002 10:36:57.136797 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2642e71-e57e-49df-ae69-318c2e6b06ec" containerName="registry-server" Oct 02 10:36:57 crc kubenswrapper[4934]: I1002 10:36:57.137855 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-swl7r" Oct 02 10:36:57 crc kubenswrapper[4934]: I1002 10:36:57.156636 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-swl7r"] Oct 02 10:36:57 crc kubenswrapper[4934]: I1002 10:36:57.225438 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47fc332b-10ec-4971-bc0f-5d42d181cf3d-utilities\") pod \"redhat-operators-swl7r\" (UID: \"47fc332b-10ec-4971-bc0f-5d42d181cf3d\") " pod="openshift-marketplace/redhat-operators-swl7r" Oct 02 10:36:57 crc kubenswrapper[4934]: I1002 10:36:57.225613 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5d5f\" (UniqueName: \"kubernetes.io/projected/47fc332b-10ec-4971-bc0f-5d42d181cf3d-kube-api-access-f5d5f\") pod \"redhat-operators-swl7r\" (UID: \"47fc332b-10ec-4971-bc0f-5d42d181cf3d\") " pod="openshift-marketplace/redhat-operators-swl7r" Oct 02 10:36:57 crc kubenswrapper[4934]: I1002 10:36:57.225682 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47fc332b-10ec-4971-bc0f-5d42d181cf3d-catalog-content\") pod \"redhat-operators-swl7r\" (UID: \"47fc332b-10ec-4971-bc0f-5d42d181cf3d\") " pod="openshift-marketplace/redhat-operators-swl7r" Oct 02 10:36:57 crc kubenswrapper[4934]: I1002 10:36:57.326344 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47fc332b-10ec-4971-bc0f-5d42d181cf3d-utilities\") pod \"redhat-operators-swl7r\" (UID: \"47fc332b-10ec-4971-bc0f-5d42d181cf3d\") " pod="openshift-marketplace/redhat-operators-swl7r" Oct 02 10:36:57 crc kubenswrapper[4934]: I1002 10:36:57.326428 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5d5f\" (UniqueName: \"kubernetes.io/projected/47fc332b-10ec-4971-bc0f-5d42d181cf3d-kube-api-access-f5d5f\") pod \"redhat-operators-swl7r\" (UID: \"47fc332b-10ec-4971-bc0f-5d42d181cf3d\") " pod="openshift-marketplace/redhat-operators-swl7r" Oct 02 10:36:57 crc kubenswrapper[4934]: I1002 10:36:57.326456 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47fc332b-10ec-4971-bc0f-5d42d181cf3d-catalog-content\") pod \"redhat-operators-swl7r\" (UID: \"47fc332b-10ec-4971-bc0f-5d42d181cf3d\") " pod="openshift-marketplace/redhat-operators-swl7r" Oct 02 10:36:57 crc kubenswrapper[4934]: I1002 10:36:57.326928 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47fc332b-10ec-4971-bc0f-5d42d181cf3d-utilities\") pod \"redhat-operators-swl7r\" (UID: \"47fc332b-10ec-4971-bc0f-5d42d181cf3d\") " pod="openshift-marketplace/redhat-operators-swl7r" Oct 02 10:36:57 crc kubenswrapper[4934]: I1002 10:36:57.326939 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47fc332b-10ec-4971-bc0f-5d42d181cf3d-catalog-content\") pod \"redhat-operators-swl7r\" (UID: \"47fc332b-10ec-4971-bc0f-5d42d181cf3d\") " pod="openshift-marketplace/redhat-operators-swl7r" Oct 02 10:36:57 crc kubenswrapper[4934]: I1002 10:36:57.349888 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5d5f\" (UniqueName: \"kubernetes.io/projected/47fc332b-10ec-4971-bc0f-5d42d181cf3d-kube-api-access-f5d5f\") pod \"redhat-operators-swl7r\" (UID: \"47fc332b-10ec-4971-bc0f-5d42d181cf3d\") " pod="openshift-marketplace/redhat-operators-swl7r" Oct 02 10:36:57 crc kubenswrapper[4934]: I1002 10:36:57.461207 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-swl7r" Oct 02 10:36:57 crc kubenswrapper[4934]: I1002 10:36:57.924904 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-swl7r"] Oct 02 10:36:58 crc kubenswrapper[4934]: I1002 10:36:58.779268 4934 generic.go:334] "Generic (PLEG): container finished" podID="47fc332b-10ec-4971-bc0f-5d42d181cf3d" containerID="91f3a968fc1944d297a78fe1f1ac34703238ff3304051b952cb1fd450720cfdf" exitCode=0 Oct 02 10:36:58 crc kubenswrapper[4934]: I1002 10:36:58.779323 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-swl7r" event={"ID":"47fc332b-10ec-4971-bc0f-5d42d181cf3d","Type":"ContainerDied","Data":"91f3a968fc1944d297a78fe1f1ac34703238ff3304051b952cb1fd450720cfdf"} Oct 02 10:36:58 crc kubenswrapper[4934]: I1002 10:36:58.779595 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-swl7r" event={"ID":"47fc332b-10ec-4971-bc0f-5d42d181cf3d","Type":"ContainerStarted","Data":"7c0f08c38eae4df575562f9002c56210f338aa7ff9ba2fe7dd8c209e714818ff"} Oct 02 10:36:58 crc kubenswrapper[4934]: I1002 10:36:58.786204 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 10:36:59 crc kubenswrapper[4934]: I1002 10:36:59.790232 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-swl7r" event={"ID":"47fc332b-10ec-4971-bc0f-5d42d181cf3d","Type":"ContainerStarted","Data":"2182c524379012494b15fd51360edae3ac15fdff6f7557596fc490f1f272d4bd"} Oct 02 10:37:00 crc kubenswrapper[4934]: I1002 10:37:00.802462 4934 generic.go:334] "Generic (PLEG): container finished" podID="47fc332b-10ec-4971-bc0f-5d42d181cf3d" containerID="2182c524379012494b15fd51360edae3ac15fdff6f7557596fc490f1f272d4bd" exitCode=0 Oct 02 10:37:00 crc kubenswrapper[4934]: I1002 10:37:00.802601 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-swl7r" event={"ID":"47fc332b-10ec-4971-bc0f-5d42d181cf3d","Type":"ContainerDied","Data":"2182c524379012494b15fd51360edae3ac15fdff6f7557596fc490f1f272d4bd"} Oct 02 10:37:01 crc kubenswrapper[4934]: I1002 10:37:01.814331 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-swl7r" event={"ID":"47fc332b-10ec-4971-bc0f-5d42d181cf3d","Type":"ContainerStarted","Data":"b87f2460915078792bf3afbbb8147478635ea3823da05a312c34b5821514599b"} Oct 02 10:37:01 crc kubenswrapper[4934]: I1002 10:37:01.848785 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-swl7r" podStartSLOduration=2.277183487 podStartE2EDuration="4.848769342s" podCreationTimestamp="2025-10-02 10:36:57 +0000 UTC" firstStartedPulling="2025-10-02 10:36:58.785843562 +0000 UTC m=+2890.538485104" lastFinishedPulling="2025-10-02 10:37:01.357429417 +0000 UTC m=+2893.110070959" observedRunningTime="2025-10-02 10:37:01.845125242 +0000 UTC m=+2893.597766764" watchObservedRunningTime="2025-10-02 10:37:01.848769342 +0000 UTC m=+2893.601410864" Oct 02 10:37:07 crc kubenswrapper[4934]: I1002 10:37:07.462055 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-swl7r" Oct 02 10:37:07 crc kubenswrapper[4934]: I1002 10:37:07.463004 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-swl7r" Oct 02 10:37:07 crc kubenswrapper[4934]: I1002 10:37:07.510741 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-swl7r" Oct 02 10:37:07 crc kubenswrapper[4934]: I1002 10:37:07.915130 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-swl7r" Oct 02 10:37:07 crc kubenswrapper[4934]: I1002 10:37:07.976732 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-swl7r"] Oct 02 10:37:08 crc kubenswrapper[4934]: I1002 10:37:08.439558 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:37:08 crc kubenswrapper[4934]: I1002 10:37:08.440227 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:37:08 crc kubenswrapper[4934]: I1002 10:37:08.440374 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 10:37:08 crc kubenswrapper[4934]: I1002 10:37:08.441156 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5"} pod="openshift-machine-config-operator/machine-config-daemon-djh5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 10:37:08 crc kubenswrapper[4934]: I1002 10:37:08.441336 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" containerID="cri-o://4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5" gracePeriod=600 Oct 02 10:37:08 crc kubenswrapper[4934]: E1002 10:37:08.573963 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:37:08 crc kubenswrapper[4934]: I1002 10:37:08.883275 4934 generic.go:334] "Generic (PLEG): container finished" podID="71db06ef-05b0-4f58-b251-b27117a8500a" containerID="4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5" exitCode=0 Oct 02 10:37:08 crc kubenswrapper[4934]: I1002 10:37:08.883350 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerDied","Data":"4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5"} Oct 02 10:37:08 crc kubenswrapper[4934]: I1002 10:37:08.883449 4934 scope.go:117] "RemoveContainer" containerID="4bb852b9106695a7e9f562beaad217ab76737694c80e8ef423f534bbeba47953" Oct 02 10:37:08 crc kubenswrapper[4934]: I1002 10:37:08.884065 4934 scope.go:117] "RemoveContainer" containerID="4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5" Oct 02 10:37:08 crc kubenswrapper[4934]: E1002 10:37:08.884309 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:37:09 crc kubenswrapper[4934]: I1002 10:37:09.893680 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-swl7r" podUID="47fc332b-10ec-4971-bc0f-5d42d181cf3d" containerName="registry-server" containerID="cri-o://b87f2460915078792bf3afbbb8147478635ea3823da05a312c34b5821514599b" gracePeriod=2 Oct 02 10:37:11 crc kubenswrapper[4934]: I1002 10:37:11.915652 4934 generic.go:334] "Generic (PLEG): container finished" podID="47fc332b-10ec-4971-bc0f-5d42d181cf3d" containerID="b87f2460915078792bf3afbbb8147478635ea3823da05a312c34b5821514599b" exitCode=0 Oct 02 10:37:11 crc kubenswrapper[4934]: I1002 10:37:11.915735 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-swl7r" event={"ID":"47fc332b-10ec-4971-bc0f-5d42d181cf3d","Type":"ContainerDied","Data":"b87f2460915078792bf3afbbb8147478635ea3823da05a312c34b5821514599b"} Oct 02 10:37:12 crc kubenswrapper[4934]: I1002 10:37:12.232929 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-swl7r" Oct 02 10:37:12 crc kubenswrapper[4934]: I1002 10:37:12.355326 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47fc332b-10ec-4971-bc0f-5d42d181cf3d-catalog-content\") pod \"47fc332b-10ec-4971-bc0f-5d42d181cf3d\" (UID: \"47fc332b-10ec-4971-bc0f-5d42d181cf3d\") " Oct 02 10:37:12 crc kubenswrapper[4934]: I1002 10:37:12.355426 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5d5f\" (UniqueName: \"kubernetes.io/projected/47fc332b-10ec-4971-bc0f-5d42d181cf3d-kube-api-access-f5d5f\") pod \"47fc332b-10ec-4971-bc0f-5d42d181cf3d\" (UID: \"47fc332b-10ec-4971-bc0f-5d42d181cf3d\") " Oct 02 10:37:12 crc kubenswrapper[4934]: I1002 10:37:12.355547 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47fc332b-10ec-4971-bc0f-5d42d181cf3d-utilities\") pod \"47fc332b-10ec-4971-bc0f-5d42d181cf3d\" (UID: \"47fc332b-10ec-4971-bc0f-5d42d181cf3d\") " Oct 02 10:37:12 crc kubenswrapper[4934]: I1002 10:37:12.356416 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47fc332b-10ec-4971-bc0f-5d42d181cf3d-utilities" (OuterVolumeSpecName: "utilities") pod "47fc332b-10ec-4971-bc0f-5d42d181cf3d" (UID: "47fc332b-10ec-4971-bc0f-5d42d181cf3d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:37:12 crc kubenswrapper[4934]: I1002 10:37:12.368214 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47fc332b-10ec-4971-bc0f-5d42d181cf3d-kube-api-access-f5d5f" (OuterVolumeSpecName: "kube-api-access-f5d5f") pod "47fc332b-10ec-4971-bc0f-5d42d181cf3d" (UID: "47fc332b-10ec-4971-bc0f-5d42d181cf3d"). InnerVolumeSpecName "kube-api-access-f5d5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:37:12 crc kubenswrapper[4934]: I1002 10:37:12.439323 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47fc332b-10ec-4971-bc0f-5d42d181cf3d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "47fc332b-10ec-4971-bc0f-5d42d181cf3d" (UID: "47fc332b-10ec-4971-bc0f-5d42d181cf3d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:37:12 crc kubenswrapper[4934]: I1002 10:37:12.457369 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47fc332b-10ec-4971-bc0f-5d42d181cf3d-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:37:12 crc kubenswrapper[4934]: I1002 10:37:12.457413 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47fc332b-10ec-4971-bc0f-5d42d181cf3d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:37:12 crc kubenswrapper[4934]: I1002 10:37:12.457430 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5d5f\" (UniqueName: \"kubernetes.io/projected/47fc332b-10ec-4971-bc0f-5d42d181cf3d-kube-api-access-f5d5f\") on node \"crc\" DevicePath \"\"" Oct 02 10:37:12 crc kubenswrapper[4934]: I1002 10:37:12.925645 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-swl7r" event={"ID":"47fc332b-10ec-4971-bc0f-5d42d181cf3d","Type":"ContainerDied","Data":"7c0f08c38eae4df575562f9002c56210f338aa7ff9ba2fe7dd8c209e714818ff"} Oct 02 10:37:12 crc kubenswrapper[4934]: I1002 10:37:12.925683 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-swl7r" Oct 02 10:37:12 crc kubenswrapper[4934]: I1002 10:37:12.925709 4934 scope.go:117] "RemoveContainer" containerID="b87f2460915078792bf3afbbb8147478635ea3823da05a312c34b5821514599b" Oct 02 10:37:12 crc kubenswrapper[4934]: I1002 10:37:12.944060 4934 scope.go:117] "RemoveContainer" containerID="2182c524379012494b15fd51360edae3ac15fdff6f7557596fc490f1f272d4bd" Oct 02 10:37:12 crc kubenswrapper[4934]: I1002 10:37:12.959082 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-swl7r"] Oct 02 10:37:12 crc kubenswrapper[4934]: I1002 10:37:12.972718 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-swl7r"] Oct 02 10:37:12 crc kubenswrapper[4934]: I1002 10:37:12.981793 4934 scope.go:117] "RemoveContainer" containerID="91f3a968fc1944d297a78fe1f1ac34703238ff3304051b952cb1fd450720cfdf" Oct 02 10:37:14 crc kubenswrapper[4934]: I1002 10:37:14.929775 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47fc332b-10ec-4971-bc0f-5d42d181cf3d" path="/var/lib/kubelet/pods/47fc332b-10ec-4971-bc0f-5d42d181cf3d/volumes" Oct 02 10:37:23 crc kubenswrapper[4934]: I1002 10:37:23.914175 4934 scope.go:117] "RemoveContainer" containerID="4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5" Oct 02 10:37:23 crc kubenswrapper[4934]: E1002 10:37:23.915168 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:37:34 crc kubenswrapper[4934]: I1002 10:37:34.914117 4934 scope.go:117] "RemoveContainer" containerID="4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5" Oct 02 10:37:34 crc kubenswrapper[4934]: E1002 10:37:34.916851 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:37:47 crc kubenswrapper[4934]: I1002 10:37:47.913536 4934 scope.go:117] "RemoveContainer" containerID="4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5" Oct 02 10:37:47 crc kubenswrapper[4934]: E1002 10:37:47.916322 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:37:58 crc kubenswrapper[4934]: I1002 10:37:58.918158 4934 scope.go:117] "RemoveContainer" containerID="4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5" Oct 02 10:37:58 crc kubenswrapper[4934]: E1002 10:37:58.918769 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:38:13 crc kubenswrapper[4934]: I1002 10:38:13.913103 4934 scope.go:117] "RemoveContainer" containerID="4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5" Oct 02 10:38:13 crc kubenswrapper[4934]: E1002 10:38:13.915201 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:38:27 crc kubenswrapper[4934]: I1002 10:38:27.914259 4934 scope.go:117] "RemoveContainer" containerID="4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5" Oct 02 10:38:27 crc kubenswrapper[4934]: E1002 10:38:27.915362 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:38:41 crc kubenswrapper[4934]: I1002 10:38:41.915019 4934 scope.go:117] "RemoveContainer" containerID="4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5" Oct 02 10:38:41 crc kubenswrapper[4934]: E1002 10:38:41.918455 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:38:53 crc kubenswrapper[4934]: I1002 10:38:53.913228 4934 scope.go:117] "RemoveContainer" containerID="4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5" Oct 02 10:38:53 crc kubenswrapper[4934]: E1002 10:38:53.914251 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:39:06 crc kubenswrapper[4934]: I1002 10:39:06.913225 4934 scope.go:117] "RemoveContainer" containerID="4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5" Oct 02 10:39:06 crc kubenswrapper[4934]: E1002 10:39:06.914122 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:39:19 crc kubenswrapper[4934]: I1002 10:39:19.913450 4934 scope.go:117] "RemoveContainer" containerID="4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5" Oct 02 10:39:19 crc kubenswrapper[4934]: E1002 10:39:19.914316 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:39:30 crc kubenswrapper[4934]: I1002 10:39:30.913246 4934 scope.go:117] "RemoveContainer" containerID="4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5" Oct 02 10:39:30 crc kubenswrapper[4934]: E1002 10:39:30.914386 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:39:45 crc kubenswrapper[4934]: I1002 10:39:45.913447 4934 scope.go:117] "RemoveContainer" containerID="4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5" Oct 02 10:39:45 crc kubenswrapper[4934]: E1002 10:39:45.914138 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:39:57 crc kubenswrapper[4934]: I1002 10:39:57.913237 4934 scope.go:117] "RemoveContainer" containerID="4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5" Oct 02 10:39:57 crc kubenswrapper[4934]: E1002 10:39:57.914096 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:40:11 crc kubenswrapper[4934]: I1002 10:40:11.913467 4934 scope.go:117] "RemoveContainer" containerID="4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5" Oct 02 10:40:11 crc kubenswrapper[4934]: E1002 10:40:11.914931 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:40:24 crc kubenswrapper[4934]: I1002 10:40:24.913002 4934 scope.go:117] "RemoveContainer" containerID="4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5" Oct 02 10:40:24 crc kubenswrapper[4934]: E1002 10:40:24.915083 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:40:35 crc kubenswrapper[4934]: I1002 10:40:35.913244 4934 scope.go:117] "RemoveContainer" containerID="4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5" Oct 02 10:40:35 crc kubenswrapper[4934]: E1002 10:40:35.914052 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:40:45 crc kubenswrapper[4934]: I1002 10:40:45.767471 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f9dwz"] Oct 02 10:40:45 crc kubenswrapper[4934]: E1002 10:40:45.770557 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47fc332b-10ec-4971-bc0f-5d42d181cf3d" containerName="extract-content" Oct 02 10:40:45 crc kubenswrapper[4934]: I1002 10:40:45.770770 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="47fc332b-10ec-4971-bc0f-5d42d181cf3d" containerName="extract-content" Oct 02 10:40:45 crc kubenswrapper[4934]: E1002 10:40:45.770956 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47fc332b-10ec-4971-bc0f-5d42d181cf3d" containerName="extract-utilities" Oct 02 10:40:45 crc kubenswrapper[4934]: I1002 10:40:45.771080 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="47fc332b-10ec-4971-bc0f-5d42d181cf3d" containerName="extract-utilities" Oct 02 10:40:45 crc kubenswrapper[4934]: E1002 10:40:45.771214 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47fc332b-10ec-4971-bc0f-5d42d181cf3d" containerName="registry-server" Oct 02 10:40:45 crc kubenswrapper[4934]: I1002 10:40:45.771330 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="47fc332b-10ec-4971-bc0f-5d42d181cf3d" containerName="registry-server" Oct 02 10:40:45 crc kubenswrapper[4934]: I1002 10:40:45.771777 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="47fc332b-10ec-4971-bc0f-5d42d181cf3d" containerName="registry-server" Oct 02 10:40:45 crc kubenswrapper[4934]: I1002 10:40:45.773609 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f9dwz" Oct 02 10:40:45 crc kubenswrapper[4934]: I1002 10:40:45.785000 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f9dwz"] Oct 02 10:40:45 crc kubenswrapper[4934]: I1002 10:40:45.829632 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fca1684c-eda4-4780-b8cd-50de605b409a-catalog-content\") pod \"redhat-marketplace-f9dwz\" (UID: \"fca1684c-eda4-4780-b8cd-50de605b409a\") " pod="openshift-marketplace/redhat-marketplace-f9dwz" Oct 02 10:40:45 crc kubenswrapper[4934]: I1002 10:40:45.829861 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v22gt\" (UniqueName: \"kubernetes.io/projected/fca1684c-eda4-4780-b8cd-50de605b409a-kube-api-access-v22gt\") pod \"redhat-marketplace-f9dwz\" (UID: \"fca1684c-eda4-4780-b8cd-50de605b409a\") " pod="openshift-marketplace/redhat-marketplace-f9dwz" Oct 02 10:40:45 crc kubenswrapper[4934]: I1002 10:40:45.830074 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fca1684c-eda4-4780-b8cd-50de605b409a-utilities\") pod \"redhat-marketplace-f9dwz\" (UID: \"fca1684c-eda4-4780-b8cd-50de605b409a\") " pod="openshift-marketplace/redhat-marketplace-f9dwz" Oct 02 10:40:45 crc kubenswrapper[4934]: I1002 10:40:45.931448 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fca1684c-eda4-4780-b8cd-50de605b409a-catalog-content\") pod \"redhat-marketplace-f9dwz\" (UID: \"fca1684c-eda4-4780-b8cd-50de605b409a\") " pod="openshift-marketplace/redhat-marketplace-f9dwz" Oct 02 10:40:45 crc kubenswrapper[4934]: I1002 10:40:45.931558 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v22gt\" (UniqueName: \"kubernetes.io/projected/fca1684c-eda4-4780-b8cd-50de605b409a-kube-api-access-v22gt\") pod \"redhat-marketplace-f9dwz\" (UID: \"fca1684c-eda4-4780-b8cd-50de605b409a\") " pod="openshift-marketplace/redhat-marketplace-f9dwz" Oct 02 10:40:45 crc kubenswrapper[4934]: I1002 10:40:45.931698 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fca1684c-eda4-4780-b8cd-50de605b409a-utilities\") pod \"redhat-marketplace-f9dwz\" (UID: \"fca1684c-eda4-4780-b8cd-50de605b409a\") " pod="openshift-marketplace/redhat-marketplace-f9dwz" Oct 02 10:40:45 crc kubenswrapper[4934]: I1002 10:40:45.932036 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fca1684c-eda4-4780-b8cd-50de605b409a-catalog-content\") pod \"redhat-marketplace-f9dwz\" (UID: \"fca1684c-eda4-4780-b8cd-50de605b409a\") " pod="openshift-marketplace/redhat-marketplace-f9dwz" Oct 02 10:40:45 crc kubenswrapper[4934]: I1002 10:40:45.932205 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fca1684c-eda4-4780-b8cd-50de605b409a-utilities\") pod \"redhat-marketplace-f9dwz\" (UID: \"fca1684c-eda4-4780-b8cd-50de605b409a\") " pod="openshift-marketplace/redhat-marketplace-f9dwz" Oct 02 10:40:45 crc kubenswrapper[4934]: I1002 10:40:45.957523 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v22gt\" (UniqueName: \"kubernetes.io/projected/fca1684c-eda4-4780-b8cd-50de605b409a-kube-api-access-v22gt\") pod \"redhat-marketplace-f9dwz\" (UID: \"fca1684c-eda4-4780-b8cd-50de605b409a\") " pod="openshift-marketplace/redhat-marketplace-f9dwz" Oct 02 10:40:46 crc kubenswrapper[4934]: I1002 10:40:46.112306 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f9dwz" Oct 02 10:40:46 crc kubenswrapper[4934]: I1002 10:40:46.389960 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f9dwz"] Oct 02 10:40:46 crc kubenswrapper[4934]: I1002 10:40:46.743472 4934 generic.go:334] "Generic (PLEG): container finished" podID="fca1684c-eda4-4780-b8cd-50de605b409a" containerID="2c990c2e576d3e9132cff7737d3947abfcabc6fe7e547c832a6b83648dd85c0c" exitCode=0 Oct 02 10:40:46 crc kubenswrapper[4934]: I1002 10:40:46.743543 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9dwz" event={"ID":"fca1684c-eda4-4780-b8cd-50de605b409a","Type":"ContainerDied","Data":"2c990c2e576d3e9132cff7737d3947abfcabc6fe7e547c832a6b83648dd85c0c"} Oct 02 10:40:46 crc kubenswrapper[4934]: I1002 10:40:46.743644 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9dwz" event={"ID":"fca1684c-eda4-4780-b8cd-50de605b409a","Type":"ContainerStarted","Data":"0d56866fd9222e2f48a6f34865275ef5722e01c743e62d7ba94db6c00bd9031c"} Oct 02 10:40:48 crc kubenswrapper[4934]: I1002 10:40:48.765331 4934 generic.go:334] "Generic (PLEG): container finished" podID="fca1684c-eda4-4780-b8cd-50de605b409a" containerID="5bd57fa14b419aba908794764e36c21032195f0179b50f44f80599338e947d3c" exitCode=0 Oct 02 10:40:48 crc kubenswrapper[4934]: I1002 10:40:48.765531 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9dwz" event={"ID":"fca1684c-eda4-4780-b8cd-50de605b409a","Type":"ContainerDied","Data":"5bd57fa14b419aba908794764e36c21032195f0179b50f44f80599338e947d3c"} Oct 02 10:40:49 crc kubenswrapper[4934]: I1002 10:40:49.777454 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9dwz" event={"ID":"fca1684c-eda4-4780-b8cd-50de605b409a","Type":"ContainerStarted","Data":"0b8c654279636ccd1998ce5e36999539f0f98a0c38bb9307b6f6ac1b5871826b"} Oct 02 10:40:49 crc kubenswrapper[4934]: I1002 10:40:49.811033 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f9dwz" podStartSLOduration=2.289049856 podStartE2EDuration="4.811002388s" podCreationTimestamp="2025-10-02 10:40:45 +0000 UTC" firstStartedPulling="2025-10-02 10:40:46.74561545 +0000 UTC m=+3118.498256972" lastFinishedPulling="2025-10-02 10:40:49.267567982 +0000 UTC m=+3121.020209504" observedRunningTime="2025-10-02 10:40:49.79912099 +0000 UTC m=+3121.551762512" watchObservedRunningTime="2025-10-02 10:40:49.811002388 +0000 UTC m=+3121.563643940" Oct 02 10:40:50 crc kubenswrapper[4934]: I1002 10:40:50.914125 4934 scope.go:117] "RemoveContainer" containerID="4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5" Oct 02 10:40:50 crc kubenswrapper[4934]: E1002 10:40:50.914527 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:40:56 crc kubenswrapper[4934]: I1002 10:40:56.113260 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f9dwz" Oct 02 10:40:56 crc kubenswrapper[4934]: I1002 10:40:56.113982 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f9dwz" Oct 02 10:40:56 crc kubenswrapper[4934]: I1002 10:40:56.199889 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f9dwz" Oct 02 10:40:56 crc kubenswrapper[4934]: I1002 10:40:56.909225 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f9dwz" Oct 02 10:40:56 crc kubenswrapper[4934]: I1002 10:40:56.960403 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f9dwz"] Oct 02 10:40:58 crc kubenswrapper[4934]: I1002 10:40:58.878981 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-f9dwz" podUID="fca1684c-eda4-4780-b8cd-50de605b409a" containerName="registry-server" containerID="cri-o://0b8c654279636ccd1998ce5e36999539f0f98a0c38bb9307b6f6ac1b5871826b" gracePeriod=2 Oct 02 10:40:59 crc kubenswrapper[4934]: I1002 10:40:59.883679 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f9dwz" Oct 02 10:40:59 crc kubenswrapper[4934]: I1002 10:40:59.888044 4934 generic.go:334] "Generic (PLEG): container finished" podID="fca1684c-eda4-4780-b8cd-50de605b409a" containerID="0b8c654279636ccd1998ce5e36999539f0f98a0c38bb9307b6f6ac1b5871826b" exitCode=0 Oct 02 10:40:59 crc kubenswrapper[4934]: I1002 10:40:59.888092 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9dwz" event={"ID":"fca1684c-eda4-4780-b8cd-50de605b409a","Type":"ContainerDied","Data":"0b8c654279636ccd1998ce5e36999539f0f98a0c38bb9307b6f6ac1b5871826b"} Oct 02 10:40:59 crc kubenswrapper[4934]: I1002 10:40:59.888116 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9dwz" event={"ID":"fca1684c-eda4-4780-b8cd-50de605b409a","Type":"ContainerDied","Data":"0d56866fd9222e2f48a6f34865275ef5722e01c743e62d7ba94db6c00bd9031c"} Oct 02 10:40:59 crc kubenswrapper[4934]: I1002 10:40:59.888120 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f9dwz" Oct 02 10:40:59 crc kubenswrapper[4934]: I1002 10:40:59.888133 4934 scope.go:117] "RemoveContainer" containerID="0b8c654279636ccd1998ce5e36999539f0f98a0c38bb9307b6f6ac1b5871826b" Oct 02 10:40:59 crc kubenswrapper[4934]: I1002 10:40:59.912028 4934 scope.go:117] "RemoveContainer" containerID="5bd57fa14b419aba908794764e36c21032195f0179b50f44f80599338e947d3c" Oct 02 10:40:59 crc kubenswrapper[4934]: I1002 10:40:59.930805 4934 scope.go:117] "RemoveContainer" containerID="2c990c2e576d3e9132cff7737d3947abfcabc6fe7e547c832a6b83648dd85c0c" Oct 02 10:40:59 crc kubenswrapper[4934]: I1002 10:40:59.953504 4934 scope.go:117] "RemoveContainer" containerID="0b8c654279636ccd1998ce5e36999539f0f98a0c38bb9307b6f6ac1b5871826b" Oct 02 10:40:59 crc kubenswrapper[4934]: E1002 10:40:59.953866 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b8c654279636ccd1998ce5e36999539f0f98a0c38bb9307b6f6ac1b5871826b\": container with ID starting with 0b8c654279636ccd1998ce5e36999539f0f98a0c38bb9307b6f6ac1b5871826b not found: ID does not exist" containerID="0b8c654279636ccd1998ce5e36999539f0f98a0c38bb9307b6f6ac1b5871826b" Oct 02 10:40:59 crc kubenswrapper[4934]: I1002 10:40:59.953895 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b8c654279636ccd1998ce5e36999539f0f98a0c38bb9307b6f6ac1b5871826b"} err="failed to get container status \"0b8c654279636ccd1998ce5e36999539f0f98a0c38bb9307b6f6ac1b5871826b\": rpc error: code = NotFound desc = could not find container \"0b8c654279636ccd1998ce5e36999539f0f98a0c38bb9307b6f6ac1b5871826b\": container with ID starting with 0b8c654279636ccd1998ce5e36999539f0f98a0c38bb9307b6f6ac1b5871826b not found: ID does not exist" Oct 02 10:40:59 crc kubenswrapper[4934]: I1002 10:40:59.953915 4934 scope.go:117] "RemoveContainer" containerID="5bd57fa14b419aba908794764e36c21032195f0179b50f44f80599338e947d3c" Oct 02 10:40:59 crc kubenswrapper[4934]: E1002 10:40:59.954658 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bd57fa14b419aba908794764e36c21032195f0179b50f44f80599338e947d3c\": container with ID starting with 5bd57fa14b419aba908794764e36c21032195f0179b50f44f80599338e947d3c not found: ID does not exist" containerID="5bd57fa14b419aba908794764e36c21032195f0179b50f44f80599338e947d3c" Oct 02 10:40:59 crc kubenswrapper[4934]: I1002 10:40:59.954705 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bd57fa14b419aba908794764e36c21032195f0179b50f44f80599338e947d3c"} err="failed to get container status \"5bd57fa14b419aba908794764e36c21032195f0179b50f44f80599338e947d3c\": rpc error: code = NotFound desc = could not find container \"5bd57fa14b419aba908794764e36c21032195f0179b50f44f80599338e947d3c\": container with ID starting with 5bd57fa14b419aba908794764e36c21032195f0179b50f44f80599338e947d3c not found: ID does not exist" Oct 02 10:40:59 crc kubenswrapper[4934]: I1002 10:40:59.954736 4934 scope.go:117] "RemoveContainer" containerID="2c990c2e576d3e9132cff7737d3947abfcabc6fe7e547c832a6b83648dd85c0c" Oct 02 10:40:59 crc kubenswrapper[4934]: E1002 10:40:59.958377 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c990c2e576d3e9132cff7737d3947abfcabc6fe7e547c832a6b83648dd85c0c\": container with ID starting with 2c990c2e576d3e9132cff7737d3947abfcabc6fe7e547c832a6b83648dd85c0c not found: ID does not exist" containerID="2c990c2e576d3e9132cff7737d3947abfcabc6fe7e547c832a6b83648dd85c0c" Oct 02 10:40:59 crc kubenswrapper[4934]: I1002 10:40:59.958434 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c990c2e576d3e9132cff7737d3947abfcabc6fe7e547c832a6b83648dd85c0c"} err="failed to get container status \"2c990c2e576d3e9132cff7737d3947abfcabc6fe7e547c832a6b83648dd85c0c\": rpc error: code = NotFound desc = could not find container \"2c990c2e576d3e9132cff7737d3947abfcabc6fe7e547c832a6b83648dd85c0c\": container with ID starting with 2c990c2e576d3e9132cff7737d3947abfcabc6fe7e547c832a6b83648dd85c0c not found: ID does not exist" Oct 02 10:41:00 crc kubenswrapper[4934]: I1002 10:41:00.047401 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v22gt\" (UniqueName: \"kubernetes.io/projected/fca1684c-eda4-4780-b8cd-50de605b409a-kube-api-access-v22gt\") pod \"fca1684c-eda4-4780-b8cd-50de605b409a\" (UID: \"fca1684c-eda4-4780-b8cd-50de605b409a\") " Oct 02 10:41:00 crc kubenswrapper[4934]: I1002 10:41:00.047465 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fca1684c-eda4-4780-b8cd-50de605b409a-utilities\") pod \"fca1684c-eda4-4780-b8cd-50de605b409a\" (UID: \"fca1684c-eda4-4780-b8cd-50de605b409a\") " Oct 02 10:41:00 crc kubenswrapper[4934]: I1002 10:41:00.047504 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fca1684c-eda4-4780-b8cd-50de605b409a-catalog-content\") pod \"fca1684c-eda4-4780-b8cd-50de605b409a\" (UID: \"fca1684c-eda4-4780-b8cd-50de605b409a\") " Oct 02 10:41:00 crc kubenswrapper[4934]: I1002 10:41:00.048432 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fca1684c-eda4-4780-b8cd-50de605b409a-utilities" (OuterVolumeSpecName: "utilities") pod "fca1684c-eda4-4780-b8cd-50de605b409a" (UID: "fca1684c-eda4-4780-b8cd-50de605b409a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:41:00 crc kubenswrapper[4934]: I1002 10:41:00.048825 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fca1684c-eda4-4780-b8cd-50de605b409a-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:41:00 crc kubenswrapper[4934]: I1002 10:41:00.059189 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fca1684c-eda4-4780-b8cd-50de605b409a-kube-api-access-v22gt" (OuterVolumeSpecName: "kube-api-access-v22gt") pod "fca1684c-eda4-4780-b8cd-50de605b409a" (UID: "fca1684c-eda4-4780-b8cd-50de605b409a"). InnerVolumeSpecName "kube-api-access-v22gt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:41:00 crc kubenswrapper[4934]: I1002 10:41:00.059910 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fca1684c-eda4-4780-b8cd-50de605b409a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fca1684c-eda4-4780-b8cd-50de605b409a" (UID: "fca1684c-eda4-4780-b8cd-50de605b409a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:41:00 crc kubenswrapper[4934]: I1002 10:41:00.149988 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v22gt\" (UniqueName: \"kubernetes.io/projected/fca1684c-eda4-4780-b8cd-50de605b409a-kube-api-access-v22gt\") on node \"crc\" DevicePath \"\"" Oct 02 10:41:00 crc kubenswrapper[4934]: I1002 10:41:00.150019 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fca1684c-eda4-4780-b8cd-50de605b409a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:41:00 crc kubenswrapper[4934]: I1002 10:41:00.217258 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f9dwz"] Oct 02 10:41:00 crc kubenswrapper[4934]: I1002 10:41:00.221600 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-f9dwz"] Oct 02 10:41:00 crc kubenswrapper[4934]: I1002 10:41:00.927662 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fca1684c-eda4-4780-b8cd-50de605b409a" path="/var/lib/kubelet/pods/fca1684c-eda4-4780-b8cd-50de605b409a/volumes" Oct 02 10:41:03 crc kubenswrapper[4934]: I1002 10:41:03.913831 4934 scope.go:117] "RemoveContainer" containerID="4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5" Oct 02 10:41:03 crc kubenswrapper[4934]: E1002 10:41:03.914540 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:41:15 crc kubenswrapper[4934]: I1002 10:41:15.913286 4934 scope.go:117] "RemoveContainer" containerID="4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5" Oct 02 10:41:15 crc kubenswrapper[4934]: E1002 10:41:15.914314 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:41:28 crc kubenswrapper[4934]: I1002 10:41:28.921731 4934 scope.go:117] "RemoveContainer" containerID="4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5" Oct 02 10:41:28 crc kubenswrapper[4934]: E1002 10:41:28.922547 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:41:39 crc kubenswrapper[4934]: I1002 10:41:39.913535 4934 scope.go:117] "RemoveContainer" containerID="4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5" Oct 02 10:41:39 crc kubenswrapper[4934]: E1002 10:41:39.914367 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:41:52 crc kubenswrapper[4934]: I1002 10:41:52.914077 4934 scope.go:117] "RemoveContainer" containerID="4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5" Oct 02 10:41:52 crc kubenswrapper[4934]: E1002 10:41:52.915004 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:42:07 crc kubenswrapper[4934]: I1002 10:42:07.914392 4934 scope.go:117] "RemoveContainer" containerID="4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5" Oct 02 10:42:07 crc kubenswrapper[4934]: E1002 10:42:07.915472 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:42:20 crc kubenswrapper[4934]: I1002 10:42:20.913463 4934 scope.go:117] "RemoveContainer" containerID="4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5" Oct 02 10:42:21 crc kubenswrapper[4934]: I1002 10:42:21.619516 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerStarted","Data":"6e65f5d855a69db84d01cc034695cad0d083c86b8bb538609fe0174fad30a308"} Oct 02 10:42:51 crc kubenswrapper[4934]: I1002 10:42:51.243871 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7gt25"] Oct 02 10:42:51 crc kubenswrapper[4934]: E1002 10:42:51.244963 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fca1684c-eda4-4780-b8cd-50de605b409a" containerName="registry-server" Oct 02 10:42:51 crc kubenswrapper[4934]: I1002 10:42:51.244985 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="fca1684c-eda4-4780-b8cd-50de605b409a" containerName="registry-server" Oct 02 10:42:51 crc kubenswrapper[4934]: E1002 10:42:51.245046 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fca1684c-eda4-4780-b8cd-50de605b409a" containerName="extract-content" Oct 02 10:42:51 crc kubenswrapper[4934]: I1002 10:42:51.245056 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="fca1684c-eda4-4780-b8cd-50de605b409a" containerName="extract-content" Oct 02 10:42:51 crc kubenswrapper[4934]: E1002 10:42:51.245071 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fca1684c-eda4-4780-b8cd-50de605b409a" containerName="extract-utilities" Oct 02 10:42:51 crc kubenswrapper[4934]: I1002 10:42:51.245080 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="fca1684c-eda4-4780-b8cd-50de605b409a" containerName="extract-utilities" Oct 02 10:42:51 crc kubenswrapper[4934]: I1002 10:42:51.245294 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="fca1684c-eda4-4780-b8cd-50de605b409a" containerName="registry-server" Oct 02 10:42:51 crc kubenswrapper[4934]: I1002 10:42:51.246730 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7gt25" Oct 02 10:42:51 crc kubenswrapper[4934]: I1002 10:42:51.253171 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7gt25"] Oct 02 10:42:51 crc kubenswrapper[4934]: I1002 10:42:51.416609 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40582bc0-bb78-4c91-9e13-5fadc70d8f33-utilities\") pod \"community-operators-7gt25\" (UID: \"40582bc0-bb78-4c91-9e13-5fadc70d8f33\") " pod="openshift-marketplace/community-operators-7gt25" Oct 02 10:42:51 crc kubenswrapper[4934]: I1002 10:42:51.416705 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40582bc0-bb78-4c91-9e13-5fadc70d8f33-catalog-content\") pod \"community-operators-7gt25\" (UID: \"40582bc0-bb78-4c91-9e13-5fadc70d8f33\") " pod="openshift-marketplace/community-operators-7gt25" Oct 02 10:42:51 crc kubenswrapper[4934]: I1002 10:42:51.416747 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7txv\" (UniqueName: \"kubernetes.io/projected/40582bc0-bb78-4c91-9e13-5fadc70d8f33-kube-api-access-q7txv\") pod \"community-operators-7gt25\" (UID: \"40582bc0-bb78-4c91-9e13-5fadc70d8f33\") " pod="openshift-marketplace/community-operators-7gt25" Oct 02 10:42:51 crc kubenswrapper[4934]: I1002 10:42:51.518329 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40582bc0-bb78-4c91-9e13-5fadc70d8f33-utilities\") pod \"community-operators-7gt25\" (UID: \"40582bc0-bb78-4c91-9e13-5fadc70d8f33\") " pod="openshift-marketplace/community-operators-7gt25" Oct 02 10:42:51 crc kubenswrapper[4934]: I1002 10:42:51.518417 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40582bc0-bb78-4c91-9e13-5fadc70d8f33-catalog-content\") pod \"community-operators-7gt25\" (UID: \"40582bc0-bb78-4c91-9e13-5fadc70d8f33\") " pod="openshift-marketplace/community-operators-7gt25" Oct 02 10:42:51 crc kubenswrapper[4934]: I1002 10:42:51.518456 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7txv\" (UniqueName: \"kubernetes.io/projected/40582bc0-bb78-4c91-9e13-5fadc70d8f33-kube-api-access-q7txv\") pod \"community-operators-7gt25\" (UID: \"40582bc0-bb78-4c91-9e13-5fadc70d8f33\") " pod="openshift-marketplace/community-operators-7gt25" Oct 02 10:42:51 crc kubenswrapper[4934]: I1002 10:42:51.518921 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40582bc0-bb78-4c91-9e13-5fadc70d8f33-utilities\") pod \"community-operators-7gt25\" (UID: \"40582bc0-bb78-4c91-9e13-5fadc70d8f33\") " pod="openshift-marketplace/community-operators-7gt25" Oct 02 10:42:51 crc kubenswrapper[4934]: I1002 10:42:51.518996 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40582bc0-bb78-4c91-9e13-5fadc70d8f33-catalog-content\") pod \"community-operators-7gt25\" (UID: \"40582bc0-bb78-4c91-9e13-5fadc70d8f33\") " pod="openshift-marketplace/community-operators-7gt25" Oct 02 10:42:51 crc kubenswrapper[4934]: I1002 10:42:51.544921 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7txv\" (UniqueName: \"kubernetes.io/projected/40582bc0-bb78-4c91-9e13-5fadc70d8f33-kube-api-access-q7txv\") pod \"community-operators-7gt25\" (UID: \"40582bc0-bb78-4c91-9e13-5fadc70d8f33\") " pod="openshift-marketplace/community-operators-7gt25" Oct 02 10:42:51 crc kubenswrapper[4934]: I1002 10:42:51.568524 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7gt25" Oct 02 10:42:52 crc kubenswrapper[4934]: I1002 10:42:52.053818 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7gt25"] Oct 02 10:42:52 crc kubenswrapper[4934]: E1002 10:42:52.403861 4934 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40582bc0_bb78_4c91_9e13_5fadc70d8f33.slice/crio-conmon-9479622bcd20258ec7d93e1a7d47e357ebe3d4c132390d0ae13ec4b929536788.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40582bc0_bb78_4c91_9e13_5fadc70d8f33.slice/crio-9479622bcd20258ec7d93e1a7d47e357ebe3d4c132390d0ae13ec4b929536788.scope\": RecentStats: unable to find data in memory cache]" Oct 02 10:42:52 crc kubenswrapper[4934]: I1002 10:42:52.928893 4934 generic.go:334] "Generic (PLEG): container finished" podID="40582bc0-bb78-4c91-9e13-5fadc70d8f33" containerID="9479622bcd20258ec7d93e1a7d47e357ebe3d4c132390d0ae13ec4b929536788" exitCode=0 Oct 02 10:42:52 crc kubenswrapper[4934]: I1002 10:42:52.928943 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gt25" event={"ID":"40582bc0-bb78-4c91-9e13-5fadc70d8f33","Type":"ContainerDied","Data":"9479622bcd20258ec7d93e1a7d47e357ebe3d4c132390d0ae13ec4b929536788"} Oct 02 10:42:52 crc kubenswrapper[4934]: I1002 10:42:52.928970 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gt25" event={"ID":"40582bc0-bb78-4c91-9e13-5fadc70d8f33","Type":"ContainerStarted","Data":"8959fde5a472b979b16aab0894141ad0ac724cd4ceedaa361108f4c2b2ec4fa6"} Oct 02 10:42:52 crc kubenswrapper[4934]: I1002 10:42:52.931043 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 10:42:53 crc kubenswrapper[4934]: I1002 10:42:53.941764 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gt25" event={"ID":"40582bc0-bb78-4c91-9e13-5fadc70d8f33","Type":"ContainerStarted","Data":"ff803add036f1ae46d761d39159c3556bed56c78c60dbd28534229576e8b8073"} Oct 02 10:42:54 crc kubenswrapper[4934]: I1002 10:42:54.954440 4934 generic.go:334] "Generic (PLEG): container finished" podID="40582bc0-bb78-4c91-9e13-5fadc70d8f33" containerID="ff803add036f1ae46d761d39159c3556bed56c78c60dbd28534229576e8b8073" exitCode=0 Oct 02 10:42:54 crc kubenswrapper[4934]: I1002 10:42:54.954492 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gt25" event={"ID":"40582bc0-bb78-4c91-9e13-5fadc70d8f33","Type":"ContainerDied","Data":"ff803add036f1ae46d761d39159c3556bed56c78c60dbd28534229576e8b8073"} Oct 02 10:42:55 crc kubenswrapper[4934]: I1002 10:42:55.966062 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gt25" event={"ID":"40582bc0-bb78-4c91-9e13-5fadc70d8f33","Type":"ContainerStarted","Data":"c5b646a1a53e20cccf6266efc2b4e3499cf732fdc4e6d2d02c5ae565d562901c"} Oct 02 10:42:56 crc kubenswrapper[4934]: I1002 10:42:56.003377 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7gt25" podStartSLOduration=2.428152018 podStartE2EDuration="5.003316842s" podCreationTimestamp="2025-10-02 10:42:51 +0000 UTC" firstStartedPulling="2025-10-02 10:42:52.930648902 +0000 UTC m=+3244.683290424" lastFinishedPulling="2025-10-02 10:42:55.505813726 +0000 UTC m=+3247.258455248" observedRunningTime="2025-10-02 10:42:55.993445389 +0000 UTC m=+3247.746086961" watchObservedRunningTime="2025-10-02 10:42:56.003316842 +0000 UTC m=+3247.755958394" Oct 02 10:43:01 crc kubenswrapper[4934]: I1002 10:43:01.569946 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7gt25" Oct 02 10:43:01 crc kubenswrapper[4934]: I1002 10:43:01.570837 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7gt25" Oct 02 10:43:01 crc kubenswrapper[4934]: I1002 10:43:01.636259 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7gt25" Oct 02 10:43:02 crc kubenswrapper[4934]: I1002 10:43:02.084323 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7gt25" Oct 02 10:43:02 crc kubenswrapper[4934]: I1002 10:43:02.138112 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7gt25"] Oct 02 10:43:04 crc kubenswrapper[4934]: I1002 10:43:04.029173 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7gt25" podUID="40582bc0-bb78-4c91-9e13-5fadc70d8f33" containerName="registry-server" containerID="cri-o://c5b646a1a53e20cccf6266efc2b4e3499cf732fdc4e6d2d02c5ae565d562901c" gracePeriod=2 Oct 02 10:43:04 crc kubenswrapper[4934]: I1002 10:43:04.555284 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7gt25" Oct 02 10:43:04 crc kubenswrapper[4934]: I1002 10:43:04.751985 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40582bc0-bb78-4c91-9e13-5fadc70d8f33-catalog-content\") pod \"40582bc0-bb78-4c91-9e13-5fadc70d8f33\" (UID: \"40582bc0-bb78-4c91-9e13-5fadc70d8f33\") " Oct 02 10:43:04 crc kubenswrapper[4934]: I1002 10:43:04.752053 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7txv\" (UniqueName: \"kubernetes.io/projected/40582bc0-bb78-4c91-9e13-5fadc70d8f33-kube-api-access-q7txv\") pod \"40582bc0-bb78-4c91-9e13-5fadc70d8f33\" (UID: \"40582bc0-bb78-4c91-9e13-5fadc70d8f33\") " Oct 02 10:43:04 crc kubenswrapper[4934]: I1002 10:43:04.752090 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40582bc0-bb78-4c91-9e13-5fadc70d8f33-utilities\") pod \"40582bc0-bb78-4c91-9e13-5fadc70d8f33\" (UID: \"40582bc0-bb78-4c91-9e13-5fadc70d8f33\") " Oct 02 10:43:04 crc kubenswrapper[4934]: I1002 10:43:04.752907 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40582bc0-bb78-4c91-9e13-5fadc70d8f33-utilities" (OuterVolumeSpecName: "utilities") pod "40582bc0-bb78-4c91-9e13-5fadc70d8f33" (UID: "40582bc0-bb78-4c91-9e13-5fadc70d8f33"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:43:04 crc kubenswrapper[4934]: I1002 10:43:04.759097 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40582bc0-bb78-4c91-9e13-5fadc70d8f33-kube-api-access-q7txv" (OuterVolumeSpecName: "kube-api-access-q7txv") pod "40582bc0-bb78-4c91-9e13-5fadc70d8f33" (UID: "40582bc0-bb78-4c91-9e13-5fadc70d8f33"). InnerVolumeSpecName "kube-api-access-q7txv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:43:04 crc kubenswrapper[4934]: I1002 10:43:04.853858 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7txv\" (UniqueName: \"kubernetes.io/projected/40582bc0-bb78-4c91-9e13-5fadc70d8f33-kube-api-access-q7txv\") on node \"crc\" DevicePath \"\"" Oct 02 10:43:04 crc kubenswrapper[4934]: I1002 10:43:04.853917 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40582bc0-bb78-4c91-9e13-5fadc70d8f33-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:43:05 crc kubenswrapper[4934]: I1002 10:43:05.040879 4934 generic.go:334] "Generic (PLEG): container finished" podID="40582bc0-bb78-4c91-9e13-5fadc70d8f33" containerID="c5b646a1a53e20cccf6266efc2b4e3499cf732fdc4e6d2d02c5ae565d562901c" exitCode=0 Oct 02 10:43:05 crc kubenswrapper[4934]: I1002 10:43:05.040930 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gt25" event={"ID":"40582bc0-bb78-4c91-9e13-5fadc70d8f33","Type":"ContainerDied","Data":"c5b646a1a53e20cccf6266efc2b4e3499cf732fdc4e6d2d02c5ae565d562901c"} Oct 02 10:43:05 crc kubenswrapper[4934]: I1002 10:43:05.040968 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7gt25" event={"ID":"40582bc0-bb78-4c91-9e13-5fadc70d8f33","Type":"ContainerDied","Data":"8959fde5a472b979b16aab0894141ad0ac724cd4ceedaa361108f4c2b2ec4fa6"} Oct 02 10:43:05 crc kubenswrapper[4934]: I1002 10:43:05.041002 4934 scope.go:117] "RemoveContainer" containerID="c5b646a1a53e20cccf6266efc2b4e3499cf732fdc4e6d2d02c5ae565d562901c" Oct 02 10:43:05 crc kubenswrapper[4934]: I1002 10:43:05.041765 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7gt25" Oct 02 10:43:05 crc kubenswrapper[4934]: I1002 10:43:05.067922 4934 scope.go:117] "RemoveContainer" containerID="ff803add036f1ae46d761d39159c3556bed56c78c60dbd28534229576e8b8073" Oct 02 10:43:05 crc kubenswrapper[4934]: I1002 10:43:05.094181 4934 scope.go:117] "RemoveContainer" containerID="9479622bcd20258ec7d93e1a7d47e357ebe3d4c132390d0ae13ec4b929536788" Oct 02 10:43:05 crc kubenswrapper[4934]: I1002 10:43:05.139564 4934 scope.go:117] "RemoveContainer" containerID="c5b646a1a53e20cccf6266efc2b4e3499cf732fdc4e6d2d02c5ae565d562901c" Oct 02 10:43:05 crc kubenswrapper[4934]: E1002 10:43:05.140874 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5b646a1a53e20cccf6266efc2b4e3499cf732fdc4e6d2d02c5ae565d562901c\": container with ID starting with c5b646a1a53e20cccf6266efc2b4e3499cf732fdc4e6d2d02c5ae565d562901c not found: ID does not exist" containerID="c5b646a1a53e20cccf6266efc2b4e3499cf732fdc4e6d2d02c5ae565d562901c" Oct 02 10:43:05 crc kubenswrapper[4934]: I1002 10:43:05.140930 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5b646a1a53e20cccf6266efc2b4e3499cf732fdc4e6d2d02c5ae565d562901c"} err="failed to get container status \"c5b646a1a53e20cccf6266efc2b4e3499cf732fdc4e6d2d02c5ae565d562901c\": rpc error: code = NotFound desc = could not find container \"c5b646a1a53e20cccf6266efc2b4e3499cf732fdc4e6d2d02c5ae565d562901c\": container with ID starting with c5b646a1a53e20cccf6266efc2b4e3499cf732fdc4e6d2d02c5ae565d562901c not found: ID does not exist" Oct 02 10:43:05 crc kubenswrapper[4934]: I1002 10:43:05.140969 4934 scope.go:117] "RemoveContainer" containerID="ff803add036f1ae46d761d39159c3556bed56c78c60dbd28534229576e8b8073" Oct 02 10:43:05 crc kubenswrapper[4934]: E1002 10:43:05.142179 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff803add036f1ae46d761d39159c3556bed56c78c60dbd28534229576e8b8073\": container with ID starting with ff803add036f1ae46d761d39159c3556bed56c78c60dbd28534229576e8b8073 not found: ID does not exist" containerID="ff803add036f1ae46d761d39159c3556bed56c78c60dbd28534229576e8b8073" Oct 02 10:43:05 crc kubenswrapper[4934]: I1002 10:43:05.142225 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff803add036f1ae46d761d39159c3556bed56c78c60dbd28534229576e8b8073"} err="failed to get container status \"ff803add036f1ae46d761d39159c3556bed56c78c60dbd28534229576e8b8073\": rpc error: code = NotFound desc = could not find container \"ff803add036f1ae46d761d39159c3556bed56c78c60dbd28534229576e8b8073\": container with ID starting with ff803add036f1ae46d761d39159c3556bed56c78c60dbd28534229576e8b8073 not found: ID does not exist" Oct 02 10:43:05 crc kubenswrapper[4934]: I1002 10:43:05.142252 4934 scope.go:117] "RemoveContainer" containerID="9479622bcd20258ec7d93e1a7d47e357ebe3d4c132390d0ae13ec4b929536788" Oct 02 10:43:05 crc kubenswrapper[4934]: E1002 10:43:05.143268 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9479622bcd20258ec7d93e1a7d47e357ebe3d4c132390d0ae13ec4b929536788\": container with ID starting with 9479622bcd20258ec7d93e1a7d47e357ebe3d4c132390d0ae13ec4b929536788 not found: ID does not exist" containerID="9479622bcd20258ec7d93e1a7d47e357ebe3d4c132390d0ae13ec4b929536788" Oct 02 10:43:05 crc kubenswrapper[4934]: I1002 10:43:05.143325 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9479622bcd20258ec7d93e1a7d47e357ebe3d4c132390d0ae13ec4b929536788"} err="failed to get container status \"9479622bcd20258ec7d93e1a7d47e357ebe3d4c132390d0ae13ec4b929536788\": rpc error: code = NotFound desc = could not find container \"9479622bcd20258ec7d93e1a7d47e357ebe3d4c132390d0ae13ec4b929536788\": container with ID starting with 9479622bcd20258ec7d93e1a7d47e357ebe3d4c132390d0ae13ec4b929536788 not found: ID does not exist" Oct 02 10:43:05 crc kubenswrapper[4934]: I1002 10:43:05.201788 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40582bc0-bb78-4c91-9e13-5fadc70d8f33-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "40582bc0-bb78-4c91-9e13-5fadc70d8f33" (UID: "40582bc0-bb78-4c91-9e13-5fadc70d8f33"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:43:05 crc kubenswrapper[4934]: I1002 10:43:05.259887 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40582bc0-bb78-4c91-9e13-5fadc70d8f33-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:43:05 crc kubenswrapper[4934]: I1002 10:43:05.385174 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7gt25"] Oct 02 10:43:05 crc kubenswrapper[4934]: I1002 10:43:05.392023 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7gt25"] Oct 02 10:43:06 crc kubenswrapper[4934]: I1002 10:43:06.929877 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40582bc0-bb78-4c91-9e13-5fadc70d8f33" path="/var/lib/kubelet/pods/40582bc0-bb78-4c91-9e13-5fadc70d8f33/volumes" Oct 02 10:44:38 crc kubenswrapper[4934]: I1002 10:44:38.439433 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:44:38 crc kubenswrapper[4934]: I1002 10:44:38.440130 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:45:00 crc kubenswrapper[4934]: I1002 10:45:00.194204 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323365-nqwv5"] Oct 02 10:45:00 crc kubenswrapper[4934]: E1002 10:45:00.195270 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40582bc0-bb78-4c91-9e13-5fadc70d8f33" containerName="registry-server" Oct 02 10:45:00 crc kubenswrapper[4934]: I1002 10:45:00.195294 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="40582bc0-bb78-4c91-9e13-5fadc70d8f33" containerName="registry-server" Oct 02 10:45:00 crc kubenswrapper[4934]: E1002 10:45:00.195321 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40582bc0-bb78-4c91-9e13-5fadc70d8f33" containerName="extract-content" Oct 02 10:45:00 crc kubenswrapper[4934]: I1002 10:45:00.195333 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="40582bc0-bb78-4c91-9e13-5fadc70d8f33" containerName="extract-content" Oct 02 10:45:00 crc kubenswrapper[4934]: E1002 10:45:00.195359 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40582bc0-bb78-4c91-9e13-5fadc70d8f33" containerName="extract-utilities" Oct 02 10:45:00 crc kubenswrapper[4934]: I1002 10:45:00.195371 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="40582bc0-bb78-4c91-9e13-5fadc70d8f33" containerName="extract-utilities" Oct 02 10:45:00 crc kubenswrapper[4934]: I1002 10:45:00.195653 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="40582bc0-bb78-4c91-9e13-5fadc70d8f33" containerName="registry-server" Oct 02 10:45:00 crc kubenswrapper[4934]: I1002 10:45:00.196430 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-nqwv5" Oct 02 10:45:00 crc kubenswrapper[4934]: I1002 10:45:00.198901 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 10:45:00 crc kubenswrapper[4934]: I1002 10:45:00.199108 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 10:45:00 crc kubenswrapper[4934]: I1002 10:45:00.208541 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323365-nqwv5"] Oct 02 10:45:00 crc kubenswrapper[4934]: I1002 10:45:00.238486 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9q2hk\" (UniqueName: \"kubernetes.io/projected/31da1e61-9429-4b9e-a234-ace1a6169799-kube-api-access-9q2hk\") pod \"collect-profiles-29323365-nqwv5\" (UID: \"31da1e61-9429-4b9e-a234-ace1a6169799\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-nqwv5" Oct 02 10:45:00 crc kubenswrapper[4934]: I1002 10:45:00.238546 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31da1e61-9429-4b9e-a234-ace1a6169799-config-volume\") pod \"collect-profiles-29323365-nqwv5\" (UID: \"31da1e61-9429-4b9e-a234-ace1a6169799\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-nqwv5" Oct 02 10:45:00 crc kubenswrapper[4934]: I1002 10:45:00.238626 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31da1e61-9429-4b9e-a234-ace1a6169799-secret-volume\") pod \"collect-profiles-29323365-nqwv5\" (UID: \"31da1e61-9429-4b9e-a234-ace1a6169799\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-nqwv5" Oct 02 10:45:00 crc kubenswrapper[4934]: I1002 10:45:00.340106 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9q2hk\" (UniqueName: \"kubernetes.io/projected/31da1e61-9429-4b9e-a234-ace1a6169799-kube-api-access-9q2hk\") pod \"collect-profiles-29323365-nqwv5\" (UID: \"31da1e61-9429-4b9e-a234-ace1a6169799\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-nqwv5" Oct 02 10:45:00 crc kubenswrapper[4934]: I1002 10:45:00.340177 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31da1e61-9429-4b9e-a234-ace1a6169799-config-volume\") pod \"collect-profiles-29323365-nqwv5\" (UID: \"31da1e61-9429-4b9e-a234-ace1a6169799\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-nqwv5" Oct 02 10:45:00 crc kubenswrapper[4934]: I1002 10:45:00.340242 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31da1e61-9429-4b9e-a234-ace1a6169799-secret-volume\") pod \"collect-profiles-29323365-nqwv5\" (UID: \"31da1e61-9429-4b9e-a234-ace1a6169799\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-nqwv5" Oct 02 10:45:00 crc kubenswrapper[4934]: I1002 10:45:00.341204 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31da1e61-9429-4b9e-a234-ace1a6169799-config-volume\") pod \"collect-profiles-29323365-nqwv5\" (UID: \"31da1e61-9429-4b9e-a234-ace1a6169799\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-nqwv5" Oct 02 10:45:00 crc kubenswrapper[4934]: I1002 10:45:00.351691 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31da1e61-9429-4b9e-a234-ace1a6169799-secret-volume\") pod \"collect-profiles-29323365-nqwv5\" (UID: \"31da1e61-9429-4b9e-a234-ace1a6169799\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-nqwv5" Oct 02 10:45:00 crc kubenswrapper[4934]: I1002 10:45:00.355960 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9q2hk\" (UniqueName: \"kubernetes.io/projected/31da1e61-9429-4b9e-a234-ace1a6169799-kube-api-access-9q2hk\") pod \"collect-profiles-29323365-nqwv5\" (UID: \"31da1e61-9429-4b9e-a234-ace1a6169799\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-nqwv5" Oct 02 10:45:00 crc kubenswrapper[4934]: I1002 10:45:00.513724 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-nqwv5" Oct 02 10:45:00 crc kubenswrapper[4934]: I1002 10:45:00.939320 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323365-nqwv5"] Oct 02 10:45:01 crc kubenswrapper[4934]: I1002 10:45:01.020107 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-nqwv5" event={"ID":"31da1e61-9429-4b9e-a234-ace1a6169799","Type":"ContainerStarted","Data":"771bd7f6afd01deaa7ce2040a2b931c38d3f02aadc0ee04f32d236c4c111f027"} Oct 02 10:45:02 crc kubenswrapper[4934]: I1002 10:45:02.031752 4934 generic.go:334] "Generic (PLEG): container finished" podID="31da1e61-9429-4b9e-a234-ace1a6169799" containerID="5002123d904487b471db4245aed4124c39ba6749ef4fe950a435f7f1bc92a6ac" exitCode=0 Oct 02 10:45:02 crc kubenswrapper[4934]: I1002 10:45:02.031809 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-nqwv5" event={"ID":"31da1e61-9429-4b9e-a234-ace1a6169799","Type":"ContainerDied","Data":"5002123d904487b471db4245aed4124c39ba6749ef4fe950a435f7f1bc92a6ac"} Oct 02 10:45:03 crc kubenswrapper[4934]: I1002 10:45:03.323853 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-nqwv5" Oct 02 10:45:03 crc kubenswrapper[4934]: I1002 10:45:03.484569 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31da1e61-9429-4b9e-a234-ace1a6169799-secret-volume\") pod \"31da1e61-9429-4b9e-a234-ace1a6169799\" (UID: \"31da1e61-9429-4b9e-a234-ace1a6169799\") " Oct 02 10:45:03 crc kubenswrapper[4934]: I1002 10:45:03.484717 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9q2hk\" (UniqueName: \"kubernetes.io/projected/31da1e61-9429-4b9e-a234-ace1a6169799-kube-api-access-9q2hk\") pod \"31da1e61-9429-4b9e-a234-ace1a6169799\" (UID: \"31da1e61-9429-4b9e-a234-ace1a6169799\") " Oct 02 10:45:03 crc kubenswrapper[4934]: I1002 10:45:03.484860 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31da1e61-9429-4b9e-a234-ace1a6169799-config-volume\") pod \"31da1e61-9429-4b9e-a234-ace1a6169799\" (UID: \"31da1e61-9429-4b9e-a234-ace1a6169799\") " Oct 02 10:45:03 crc kubenswrapper[4934]: I1002 10:45:03.486280 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31da1e61-9429-4b9e-a234-ace1a6169799-config-volume" (OuterVolumeSpecName: "config-volume") pod "31da1e61-9429-4b9e-a234-ace1a6169799" (UID: "31da1e61-9429-4b9e-a234-ace1a6169799"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:45:03 crc kubenswrapper[4934]: I1002 10:45:03.492638 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31da1e61-9429-4b9e-a234-ace1a6169799-kube-api-access-9q2hk" (OuterVolumeSpecName: "kube-api-access-9q2hk") pod "31da1e61-9429-4b9e-a234-ace1a6169799" (UID: "31da1e61-9429-4b9e-a234-ace1a6169799"). InnerVolumeSpecName "kube-api-access-9q2hk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:45:03 crc kubenswrapper[4934]: I1002 10:45:03.493393 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31da1e61-9429-4b9e-a234-ace1a6169799-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "31da1e61-9429-4b9e-a234-ace1a6169799" (UID: "31da1e61-9429-4b9e-a234-ace1a6169799"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:45:03 crc kubenswrapper[4934]: I1002 10:45:03.586786 4934 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31da1e61-9429-4b9e-a234-ace1a6169799-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 10:45:03 crc kubenswrapper[4934]: I1002 10:45:03.586824 4934 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31da1e61-9429-4b9e-a234-ace1a6169799-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 10:45:03 crc kubenswrapper[4934]: I1002 10:45:03.586837 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9q2hk\" (UniqueName: \"kubernetes.io/projected/31da1e61-9429-4b9e-a234-ace1a6169799-kube-api-access-9q2hk\") on node \"crc\" DevicePath \"\"" Oct 02 10:45:04 crc kubenswrapper[4934]: I1002 10:45:04.058409 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-nqwv5" event={"ID":"31da1e61-9429-4b9e-a234-ace1a6169799","Type":"ContainerDied","Data":"771bd7f6afd01deaa7ce2040a2b931c38d3f02aadc0ee04f32d236c4c111f027"} Oct 02 10:45:04 crc kubenswrapper[4934]: I1002 10:45:04.058876 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="771bd7f6afd01deaa7ce2040a2b931c38d3f02aadc0ee04f32d236c4c111f027" Oct 02 10:45:04 crc kubenswrapper[4934]: I1002 10:45:04.058488 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323365-nqwv5" Oct 02 10:45:04 crc kubenswrapper[4934]: I1002 10:45:04.395827 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323320-cmf7d"] Oct 02 10:45:04 crc kubenswrapper[4934]: I1002 10:45:04.400087 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323320-cmf7d"] Oct 02 10:45:04 crc kubenswrapper[4934]: I1002 10:45:04.934209 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2c83244-b4c5-403e-91d0-481537b764fe" path="/var/lib/kubelet/pods/c2c83244-b4c5-403e-91d0-481537b764fe/volumes" Oct 02 10:45:08 crc kubenswrapper[4934]: I1002 10:45:08.439767 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:45:08 crc kubenswrapper[4934]: I1002 10:45:08.440113 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:45:13 crc kubenswrapper[4934]: I1002 10:45:13.722069 4934 scope.go:117] "RemoveContainer" containerID="09eb614893d7c26c569495d5413034f272a564e4ad165a4297729f1221b50851" Oct 02 10:45:38 crc kubenswrapper[4934]: I1002 10:45:38.439987 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:45:38 crc kubenswrapper[4934]: I1002 10:45:38.440535 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:45:38 crc kubenswrapper[4934]: I1002 10:45:38.440593 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 10:45:38 crc kubenswrapper[4934]: I1002 10:45:38.441111 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6e65f5d855a69db84d01cc034695cad0d083c86b8bb538609fe0174fad30a308"} pod="openshift-machine-config-operator/machine-config-daemon-djh5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 10:45:38 crc kubenswrapper[4934]: I1002 10:45:38.441161 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" containerID="cri-o://6e65f5d855a69db84d01cc034695cad0d083c86b8bb538609fe0174fad30a308" gracePeriod=600 Oct 02 10:45:39 crc kubenswrapper[4934]: I1002 10:45:39.346555 4934 generic.go:334] "Generic (PLEG): container finished" podID="71db06ef-05b0-4f58-b251-b27117a8500a" containerID="6e65f5d855a69db84d01cc034695cad0d083c86b8bb538609fe0174fad30a308" exitCode=0 Oct 02 10:45:39 crc kubenswrapper[4934]: I1002 10:45:39.346635 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerDied","Data":"6e65f5d855a69db84d01cc034695cad0d083c86b8bb538609fe0174fad30a308"} Oct 02 10:45:39 crc kubenswrapper[4934]: I1002 10:45:39.346996 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerStarted","Data":"4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d"} Oct 02 10:45:39 crc kubenswrapper[4934]: I1002 10:45:39.347024 4934 scope.go:117] "RemoveContainer" containerID="4e3342cb6a92907388f0a7c96a864c68b49c7e1de140d0acda94e0cc6740bfa5" Oct 02 10:47:38 crc kubenswrapper[4934]: I1002 10:47:38.439237 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:47:38 crc kubenswrapper[4934]: I1002 10:47:38.440211 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:47:45 crc kubenswrapper[4934]: I1002 10:47:45.301564 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-85b44"] Oct 02 10:47:45 crc kubenswrapper[4934]: E1002 10:47:45.303205 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31da1e61-9429-4b9e-a234-ace1a6169799" containerName="collect-profiles" Oct 02 10:47:45 crc kubenswrapper[4934]: I1002 10:47:45.303235 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="31da1e61-9429-4b9e-a234-ace1a6169799" containerName="collect-profiles" Oct 02 10:47:45 crc kubenswrapper[4934]: I1002 10:47:45.303503 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="31da1e61-9429-4b9e-a234-ace1a6169799" containerName="collect-profiles" Oct 02 10:47:45 crc kubenswrapper[4934]: I1002 10:47:45.306152 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-85b44" Oct 02 10:47:45 crc kubenswrapper[4934]: I1002 10:47:45.314930 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-85b44"] Oct 02 10:47:45 crc kubenswrapper[4934]: I1002 10:47:45.432203 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76c4b5f5-8263-42f8-96e8-e9c730d38330-catalog-content\") pod \"redhat-operators-85b44\" (UID: \"76c4b5f5-8263-42f8-96e8-e9c730d38330\") " pod="openshift-marketplace/redhat-operators-85b44" Oct 02 10:47:45 crc kubenswrapper[4934]: I1002 10:47:45.433023 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76c4b5f5-8263-42f8-96e8-e9c730d38330-utilities\") pod \"redhat-operators-85b44\" (UID: \"76c4b5f5-8263-42f8-96e8-e9c730d38330\") " pod="openshift-marketplace/redhat-operators-85b44" Oct 02 10:47:45 crc kubenswrapper[4934]: I1002 10:47:45.433076 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfnpm\" (UniqueName: \"kubernetes.io/projected/76c4b5f5-8263-42f8-96e8-e9c730d38330-kube-api-access-hfnpm\") pod \"redhat-operators-85b44\" (UID: \"76c4b5f5-8263-42f8-96e8-e9c730d38330\") " pod="openshift-marketplace/redhat-operators-85b44" Oct 02 10:47:45 crc kubenswrapper[4934]: I1002 10:47:45.536310 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76c4b5f5-8263-42f8-96e8-e9c730d38330-utilities\") pod \"redhat-operators-85b44\" (UID: \"76c4b5f5-8263-42f8-96e8-e9c730d38330\") " pod="openshift-marketplace/redhat-operators-85b44" Oct 02 10:47:45 crc kubenswrapper[4934]: I1002 10:47:45.536376 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfnpm\" (UniqueName: \"kubernetes.io/projected/76c4b5f5-8263-42f8-96e8-e9c730d38330-kube-api-access-hfnpm\") pod \"redhat-operators-85b44\" (UID: \"76c4b5f5-8263-42f8-96e8-e9c730d38330\") " pod="openshift-marketplace/redhat-operators-85b44" Oct 02 10:47:45 crc kubenswrapper[4934]: I1002 10:47:45.536478 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76c4b5f5-8263-42f8-96e8-e9c730d38330-catalog-content\") pod \"redhat-operators-85b44\" (UID: \"76c4b5f5-8263-42f8-96e8-e9c730d38330\") " pod="openshift-marketplace/redhat-operators-85b44" Oct 02 10:47:45 crc kubenswrapper[4934]: I1002 10:47:45.537174 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76c4b5f5-8263-42f8-96e8-e9c730d38330-catalog-content\") pod \"redhat-operators-85b44\" (UID: \"76c4b5f5-8263-42f8-96e8-e9c730d38330\") " pod="openshift-marketplace/redhat-operators-85b44" Oct 02 10:47:45 crc kubenswrapper[4934]: I1002 10:47:45.537342 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76c4b5f5-8263-42f8-96e8-e9c730d38330-utilities\") pod \"redhat-operators-85b44\" (UID: \"76c4b5f5-8263-42f8-96e8-e9c730d38330\") " pod="openshift-marketplace/redhat-operators-85b44" Oct 02 10:47:45 crc kubenswrapper[4934]: I1002 10:47:45.564145 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfnpm\" (UniqueName: \"kubernetes.io/projected/76c4b5f5-8263-42f8-96e8-e9c730d38330-kube-api-access-hfnpm\") pod \"redhat-operators-85b44\" (UID: \"76c4b5f5-8263-42f8-96e8-e9c730d38330\") " pod="openshift-marketplace/redhat-operators-85b44" Oct 02 10:47:45 crc kubenswrapper[4934]: I1002 10:47:45.628899 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-85b44" Oct 02 10:47:46 crc kubenswrapper[4934]: I1002 10:47:46.068537 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-85b44"] Oct 02 10:47:46 crc kubenswrapper[4934]: W1002 10:47:46.073747 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76c4b5f5_8263_42f8_96e8_e9c730d38330.slice/crio-afed61e5c5691b1b041deedb4b24139722a0383e65d72bae3b6dbef2acd3174f WatchSource:0}: Error finding container afed61e5c5691b1b041deedb4b24139722a0383e65d72bae3b6dbef2acd3174f: Status 404 returned error can't find the container with id afed61e5c5691b1b041deedb4b24139722a0383e65d72bae3b6dbef2acd3174f Oct 02 10:47:46 crc kubenswrapper[4934]: I1002 10:47:46.463521 4934 generic.go:334] "Generic (PLEG): container finished" podID="76c4b5f5-8263-42f8-96e8-e9c730d38330" containerID="bb25a112506fc95e3e6c98070666cfa2a3e5866c56f2f3ced3fca60be1690e58" exitCode=0 Oct 02 10:47:46 crc kubenswrapper[4934]: I1002 10:47:46.463595 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85b44" event={"ID":"76c4b5f5-8263-42f8-96e8-e9c730d38330","Type":"ContainerDied","Data":"bb25a112506fc95e3e6c98070666cfa2a3e5866c56f2f3ced3fca60be1690e58"} Oct 02 10:47:46 crc kubenswrapper[4934]: I1002 10:47:46.463841 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85b44" event={"ID":"76c4b5f5-8263-42f8-96e8-e9c730d38330","Type":"ContainerStarted","Data":"afed61e5c5691b1b041deedb4b24139722a0383e65d72bae3b6dbef2acd3174f"} Oct 02 10:47:47 crc kubenswrapper[4934]: I1002 10:47:47.294131 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wv6td"] Oct 02 10:47:47 crc kubenswrapper[4934]: I1002 10:47:47.296895 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wv6td" Oct 02 10:47:47 crc kubenswrapper[4934]: I1002 10:47:47.302432 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wv6td"] Oct 02 10:47:47 crc kubenswrapper[4934]: I1002 10:47:47.463436 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49vtn\" (UniqueName: \"kubernetes.io/projected/dc4edaab-1032-4bae-819a-3b74de5bdb00-kube-api-access-49vtn\") pod \"certified-operators-wv6td\" (UID: \"dc4edaab-1032-4bae-819a-3b74de5bdb00\") " pod="openshift-marketplace/certified-operators-wv6td" Oct 02 10:47:47 crc kubenswrapper[4934]: I1002 10:47:47.463504 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc4edaab-1032-4bae-819a-3b74de5bdb00-catalog-content\") pod \"certified-operators-wv6td\" (UID: \"dc4edaab-1032-4bae-819a-3b74de5bdb00\") " pod="openshift-marketplace/certified-operators-wv6td" Oct 02 10:47:47 crc kubenswrapper[4934]: I1002 10:47:47.463601 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc4edaab-1032-4bae-819a-3b74de5bdb00-utilities\") pod \"certified-operators-wv6td\" (UID: \"dc4edaab-1032-4bae-819a-3b74de5bdb00\") " pod="openshift-marketplace/certified-operators-wv6td" Oct 02 10:47:47 crc kubenswrapper[4934]: I1002 10:47:47.477166 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85b44" event={"ID":"76c4b5f5-8263-42f8-96e8-e9c730d38330","Type":"ContainerStarted","Data":"fc277d89f2dfe8364db632a7392b084f8be33264a8e7db67b704919213013497"} Oct 02 10:47:47 crc kubenswrapper[4934]: I1002 10:47:47.564937 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc4edaab-1032-4bae-819a-3b74de5bdb00-utilities\") pod \"certified-operators-wv6td\" (UID: \"dc4edaab-1032-4bae-819a-3b74de5bdb00\") " pod="openshift-marketplace/certified-operators-wv6td" Oct 02 10:47:47 crc kubenswrapper[4934]: I1002 10:47:47.565024 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49vtn\" (UniqueName: \"kubernetes.io/projected/dc4edaab-1032-4bae-819a-3b74de5bdb00-kube-api-access-49vtn\") pod \"certified-operators-wv6td\" (UID: \"dc4edaab-1032-4bae-819a-3b74de5bdb00\") " pod="openshift-marketplace/certified-operators-wv6td" Oct 02 10:47:47 crc kubenswrapper[4934]: I1002 10:47:47.565048 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc4edaab-1032-4bae-819a-3b74de5bdb00-catalog-content\") pod \"certified-operators-wv6td\" (UID: \"dc4edaab-1032-4bae-819a-3b74de5bdb00\") " pod="openshift-marketplace/certified-operators-wv6td" Oct 02 10:47:47 crc kubenswrapper[4934]: I1002 10:47:47.565625 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc4edaab-1032-4bae-819a-3b74de5bdb00-catalog-content\") pod \"certified-operators-wv6td\" (UID: \"dc4edaab-1032-4bae-819a-3b74de5bdb00\") " pod="openshift-marketplace/certified-operators-wv6td" Oct 02 10:47:47 crc kubenswrapper[4934]: I1002 10:47:47.565631 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc4edaab-1032-4bae-819a-3b74de5bdb00-utilities\") pod \"certified-operators-wv6td\" (UID: \"dc4edaab-1032-4bae-819a-3b74de5bdb00\") " pod="openshift-marketplace/certified-operators-wv6td" Oct 02 10:47:47 crc kubenswrapper[4934]: I1002 10:47:47.588826 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49vtn\" (UniqueName: \"kubernetes.io/projected/dc4edaab-1032-4bae-819a-3b74de5bdb00-kube-api-access-49vtn\") pod \"certified-operators-wv6td\" (UID: \"dc4edaab-1032-4bae-819a-3b74de5bdb00\") " pod="openshift-marketplace/certified-operators-wv6td" Oct 02 10:47:47 crc kubenswrapper[4934]: I1002 10:47:47.650890 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wv6td" Oct 02 10:47:48 crc kubenswrapper[4934]: I1002 10:47:48.110580 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wv6td"] Oct 02 10:47:48 crc kubenswrapper[4934]: I1002 10:47:48.489384 4934 generic.go:334] "Generic (PLEG): container finished" podID="76c4b5f5-8263-42f8-96e8-e9c730d38330" containerID="fc277d89f2dfe8364db632a7392b084f8be33264a8e7db67b704919213013497" exitCode=0 Oct 02 10:47:48 crc kubenswrapper[4934]: I1002 10:47:48.489771 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85b44" event={"ID":"76c4b5f5-8263-42f8-96e8-e9c730d38330","Type":"ContainerDied","Data":"fc277d89f2dfe8364db632a7392b084f8be33264a8e7db67b704919213013497"} Oct 02 10:47:48 crc kubenswrapper[4934]: I1002 10:47:48.493709 4934 generic.go:334] "Generic (PLEG): container finished" podID="dc4edaab-1032-4bae-819a-3b74de5bdb00" containerID="b5a7f98ab9cc9993b18e3dbf8ba994517fe984ea191805f8373b0da54911f4a0" exitCode=0 Oct 02 10:47:48 crc kubenswrapper[4934]: I1002 10:47:48.493743 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wv6td" event={"ID":"dc4edaab-1032-4bae-819a-3b74de5bdb00","Type":"ContainerDied","Data":"b5a7f98ab9cc9993b18e3dbf8ba994517fe984ea191805f8373b0da54911f4a0"} Oct 02 10:47:48 crc kubenswrapper[4934]: I1002 10:47:48.493766 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wv6td" event={"ID":"dc4edaab-1032-4bae-819a-3b74de5bdb00","Type":"ContainerStarted","Data":"e2414eaaaf13f7bcdb1d834804a2a5f874487bf35887fc753717676bb78859e5"} Oct 02 10:47:49 crc kubenswrapper[4934]: I1002 10:47:49.504814 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85b44" event={"ID":"76c4b5f5-8263-42f8-96e8-e9c730d38330","Type":"ContainerStarted","Data":"0ffc734829126f15024640dfcb73f88db5dfa5618393e7eb06ddef530c7354f5"} Oct 02 10:47:49 crc kubenswrapper[4934]: I1002 10:47:49.507245 4934 generic.go:334] "Generic (PLEG): container finished" podID="dc4edaab-1032-4bae-819a-3b74de5bdb00" containerID="7ba52a88e3284066e85ee1b9e1672b6420ccd8abbbd4aa398edd6427d50222dd" exitCode=0 Oct 02 10:47:49 crc kubenswrapper[4934]: I1002 10:47:49.507280 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wv6td" event={"ID":"dc4edaab-1032-4bae-819a-3b74de5bdb00","Type":"ContainerDied","Data":"7ba52a88e3284066e85ee1b9e1672b6420ccd8abbbd4aa398edd6427d50222dd"} Oct 02 10:47:49 crc kubenswrapper[4934]: I1002 10:47:49.527426 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-85b44" podStartSLOduration=2.071551372 podStartE2EDuration="4.527405159s" podCreationTimestamp="2025-10-02 10:47:45 +0000 UTC" firstStartedPulling="2025-10-02 10:47:46.465684124 +0000 UTC m=+3538.218325646" lastFinishedPulling="2025-10-02 10:47:48.921537901 +0000 UTC m=+3540.674179433" observedRunningTime="2025-10-02 10:47:49.522468013 +0000 UTC m=+3541.275109535" watchObservedRunningTime="2025-10-02 10:47:49.527405159 +0000 UTC m=+3541.280046681" Oct 02 10:47:50 crc kubenswrapper[4934]: I1002 10:47:50.514766 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wv6td" event={"ID":"dc4edaab-1032-4bae-819a-3b74de5bdb00","Type":"ContainerStarted","Data":"a6d68ed44d2d02956eb58b9cff75e811ad2fe2d892747b73b415eb5e1d72ed16"} Oct 02 10:47:55 crc kubenswrapper[4934]: I1002 10:47:55.629906 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-85b44" Oct 02 10:47:55 crc kubenswrapper[4934]: I1002 10:47:55.630347 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-85b44" Oct 02 10:47:55 crc kubenswrapper[4934]: I1002 10:47:55.700010 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-85b44" Oct 02 10:47:55 crc kubenswrapper[4934]: I1002 10:47:55.727469 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wv6td" podStartSLOduration=7.298069899 podStartE2EDuration="8.727443334s" podCreationTimestamp="2025-10-02 10:47:47 +0000 UTC" firstStartedPulling="2025-10-02 10:47:48.495973838 +0000 UTC m=+3540.248615390" lastFinishedPulling="2025-10-02 10:47:49.925347303 +0000 UTC m=+3541.677988825" observedRunningTime="2025-10-02 10:47:50.547334303 +0000 UTC m=+3542.299975825" watchObservedRunningTime="2025-10-02 10:47:55.727443334 +0000 UTC m=+3547.480084876" Oct 02 10:47:56 crc kubenswrapper[4934]: I1002 10:47:56.647487 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-85b44" Oct 02 10:47:57 crc kubenswrapper[4934]: I1002 10:47:57.652270 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wv6td" Oct 02 10:47:57 crc kubenswrapper[4934]: I1002 10:47:57.652349 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wv6td" Oct 02 10:47:57 crc kubenswrapper[4934]: I1002 10:47:57.728055 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wv6td" Oct 02 10:47:58 crc kubenswrapper[4934]: I1002 10:47:58.672042 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wv6td" Oct 02 10:47:58 crc kubenswrapper[4934]: I1002 10:47:58.678270 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-85b44"] Oct 02 10:47:58 crc kubenswrapper[4934]: I1002 10:47:58.678563 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-85b44" podUID="76c4b5f5-8263-42f8-96e8-e9c730d38330" containerName="registry-server" containerID="cri-o://0ffc734829126f15024640dfcb73f88db5dfa5618393e7eb06ddef530c7354f5" gracePeriod=2 Oct 02 10:48:00 crc kubenswrapper[4934]: I1002 10:48:00.077475 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wv6td"] Oct 02 10:48:00 crc kubenswrapper[4934]: I1002 10:48:00.611260 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wv6td" podUID="dc4edaab-1032-4bae-819a-3b74de5bdb00" containerName="registry-server" containerID="cri-o://a6d68ed44d2d02956eb58b9cff75e811ad2fe2d892747b73b415eb5e1d72ed16" gracePeriod=2 Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.294943 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-85b44" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.474989 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76c4b5f5-8263-42f8-96e8-e9c730d38330-catalog-content\") pod \"76c4b5f5-8263-42f8-96e8-e9c730d38330\" (UID: \"76c4b5f5-8263-42f8-96e8-e9c730d38330\") " Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.475040 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfnpm\" (UniqueName: \"kubernetes.io/projected/76c4b5f5-8263-42f8-96e8-e9c730d38330-kube-api-access-hfnpm\") pod \"76c4b5f5-8263-42f8-96e8-e9c730d38330\" (UID: \"76c4b5f5-8263-42f8-96e8-e9c730d38330\") " Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.475181 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76c4b5f5-8263-42f8-96e8-e9c730d38330-utilities\") pod \"76c4b5f5-8263-42f8-96e8-e9c730d38330\" (UID: \"76c4b5f5-8263-42f8-96e8-e9c730d38330\") " Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.476311 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76c4b5f5-8263-42f8-96e8-e9c730d38330-utilities" (OuterVolumeSpecName: "utilities") pod "76c4b5f5-8263-42f8-96e8-e9c730d38330" (UID: "76c4b5f5-8263-42f8-96e8-e9c730d38330"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.481071 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76c4b5f5-8263-42f8-96e8-e9c730d38330-kube-api-access-hfnpm" (OuterVolumeSpecName: "kube-api-access-hfnpm") pod "76c4b5f5-8263-42f8-96e8-e9c730d38330" (UID: "76c4b5f5-8263-42f8-96e8-e9c730d38330"). InnerVolumeSpecName "kube-api-access-hfnpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.492820 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wv6td" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.562407 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76c4b5f5-8263-42f8-96e8-e9c730d38330-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "76c4b5f5-8263-42f8-96e8-e9c730d38330" (UID: "76c4b5f5-8263-42f8-96e8-e9c730d38330"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.577881 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfnpm\" (UniqueName: \"kubernetes.io/projected/76c4b5f5-8263-42f8-96e8-e9c730d38330-kube-api-access-hfnpm\") on node \"crc\" DevicePath \"\"" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.577925 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76c4b5f5-8263-42f8-96e8-e9c730d38330-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.577938 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76c4b5f5-8263-42f8-96e8-e9c730d38330-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.624415 4934 generic.go:334] "Generic (PLEG): container finished" podID="76c4b5f5-8263-42f8-96e8-e9c730d38330" containerID="0ffc734829126f15024640dfcb73f88db5dfa5618393e7eb06ddef530c7354f5" exitCode=0 Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.624780 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85b44" event={"ID":"76c4b5f5-8263-42f8-96e8-e9c730d38330","Type":"ContainerDied","Data":"0ffc734829126f15024640dfcb73f88db5dfa5618393e7eb06ddef530c7354f5"} Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.624857 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85b44" event={"ID":"76c4b5f5-8263-42f8-96e8-e9c730d38330","Type":"ContainerDied","Data":"afed61e5c5691b1b041deedb4b24139722a0383e65d72bae3b6dbef2acd3174f"} Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.624892 4934 scope.go:117] "RemoveContainer" containerID="0ffc734829126f15024640dfcb73f88db5dfa5618393e7eb06ddef530c7354f5" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.624805 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-85b44" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.627943 4934 generic.go:334] "Generic (PLEG): container finished" podID="dc4edaab-1032-4bae-819a-3b74de5bdb00" containerID="a6d68ed44d2d02956eb58b9cff75e811ad2fe2d892747b73b415eb5e1d72ed16" exitCode=0 Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.627991 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wv6td" event={"ID":"dc4edaab-1032-4bae-819a-3b74de5bdb00","Type":"ContainerDied","Data":"a6d68ed44d2d02956eb58b9cff75e811ad2fe2d892747b73b415eb5e1d72ed16"} Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.628013 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wv6td" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.628027 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wv6td" event={"ID":"dc4edaab-1032-4bae-819a-3b74de5bdb00","Type":"ContainerDied","Data":"e2414eaaaf13f7bcdb1d834804a2a5f874487bf35887fc753717676bb78859e5"} Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.648924 4934 scope.go:117] "RemoveContainer" containerID="fc277d89f2dfe8364db632a7392b084f8be33264a8e7db67b704919213013497" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.658030 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-85b44"] Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.662897 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-85b44"] Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.678998 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc4edaab-1032-4bae-819a-3b74de5bdb00-catalog-content\") pod \"dc4edaab-1032-4bae-819a-3b74de5bdb00\" (UID: \"dc4edaab-1032-4bae-819a-3b74de5bdb00\") " Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.679049 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc4edaab-1032-4bae-819a-3b74de5bdb00-utilities\") pod \"dc4edaab-1032-4bae-819a-3b74de5bdb00\" (UID: \"dc4edaab-1032-4bae-819a-3b74de5bdb00\") " Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.679235 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49vtn\" (UniqueName: \"kubernetes.io/projected/dc4edaab-1032-4bae-819a-3b74de5bdb00-kube-api-access-49vtn\") pod \"dc4edaab-1032-4bae-819a-3b74de5bdb00\" (UID: \"dc4edaab-1032-4bae-819a-3b74de5bdb00\") " Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.679814 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc4edaab-1032-4bae-819a-3b74de5bdb00-utilities" (OuterVolumeSpecName: "utilities") pod "dc4edaab-1032-4bae-819a-3b74de5bdb00" (UID: "dc4edaab-1032-4bae-819a-3b74de5bdb00"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.682430 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc4edaab-1032-4bae-819a-3b74de5bdb00-kube-api-access-49vtn" (OuterVolumeSpecName: "kube-api-access-49vtn") pod "dc4edaab-1032-4bae-819a-3b74de5bdb00" (UID: "dc4edaab-1032-4bae-819a-3b74de5bdb00"). InnerVolumeSpecName "kube-api-access-49vtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.689514 4934 scope.go:117] "RemoveContainer" containerID="bb25a112506fc95e3e6c98070666cfa2a3e5866c56f2f3ced3fca60be1690e58" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.704845 4934 scope.go:117] "RemoveContainer" containerID="0ffc734829126f15024640dfcb73f88db5dfa5618393e7eb06ddef530c7354f5" Oct 02 10:48:01 crc kubenswrapper[4934]: E1002 10:48:01.705444 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ffc734829126f15024640dfcb73f88db5dfa5618393e7eb06ddef530c7354f5\": container with ID starting with 0ffc734829126f15024640dfcb73f88db5dfa5618393e7eb06ddef530c7354f5 not found: ID does not exist" containerID="0ffc734829126f15024640dfcb73f88db5dfa5618393e7eb06ddef530c7354f5" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.705611 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ffc734829126f15024640dfcb73f88db5dfa5618393e7eb06ddef530c7354f5"} err="failed to get container status \"0ffc734829126f15024640dfcb73f88db5dfa5618393e7eb06ddef530c7354f5\": rpc error: code = NotFound desc = could not find container \"0ffc734829126f15024640dfcb73f88db5dfa5618393e7eb06ddef530c7354f5\": container with ID starting with 0ffc734829126f15024640dfcb73f88db5dfa5618393e7eb06ddef530c7354f5 not found: ID does not exist" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.705708 4934 scope.go:117] "RemoveContainer" containerID="fc277d89f2dfe8364db632a7392b084f8be33264a8e7db67b704919213013497" Oct 02 10:48:01 crc kubenswrapper[4934]: E1002 10:48:01.706195 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc277d89f2dfe8364db632a7392b084f8be33264a8e7db67b704919213013497\": container with ID starting with fc277d89f2dfe8364db632a7392b084f8be33264a8e7db67b704919213013497 not found: ID does not exist" containerID="fc277d89f2dfe8364db632a7392b084f8be33264a8e7db67b704919213013497" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.706223 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc277d89f2dfe8364db632a7392b084f8be33264a8e7db67b704919213013497"} err="failed to get container status \"fc277d89f2dfe8364db632a7392b084f8be33264a8e7db67b704919213013497\": rpc error: code = NotFound desc = could not find container \"fc277d89f2dfe8364db632a7392b084f8be33264a8e7db67b704919213013497\": container with ID starting with fc277d89f2dfe8364db632a7392b084f8be33264a8e7db67b704919213013497 not found: ID does not exist" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.706242 4934 scope.go:117] "RemoveContainer" containerID="bb25a112506fc95e3e6c98070666cfa2a3e5866c56f2f3ced3fca60be1690e58" Oct 02 10:48:01 crc kubenswrapper[4934]: E1002 10:48:01.706509 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb25a112506fc95e3e6c98070666cfa2a3e5866c56f2f3ced3fca60be1690e58\": container with ID starting with bb25a112506fc95e3e6c98070666cfa2a3e5866c56f2f3ced3fca60be1690e58 not found: ID does not exist" containerID="bb25a112506fc95e3e6c98070666cfa2a3e5866c56f2f3ced3fca60be1690e58" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.706737 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb25a112506fc95e3e6c98070666cfa2a3e5866c56f2f3ced3fca60be1690e58"} err="failed to get container status \"bb25a112506fc95e3e6c98070666cfa2a3e5866c56f2f3ced3fca60be1690e58\": rpc error: code = NotFound desc = could not find container \"bb25a112506fc95e3e6c98070666cfa2a3e5866c56f2f3ced3fca60be1690e58\": container with ID starting with bb25a112506fc95e3e6c98070666cfa2a3e5866c56f2f3ced3fca60be1690e58 not found: ID does not exist" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.706822 4934 scope.go:117] "RemoveContainer" containerID="a6d68ed44d2d02956eb58b9cff75e811ad2fe2d892747b73b415eb5e1d72ed16" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.722485 4934 scope.go:117] "RemoveContainer" containerID="7ba52a88e3284066e85ee1b9e1672b6420ccd8abbbd4aa398edd6427d50222dd" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.727270 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc4edaab-1032-4bae-819a-3b74de5bdb00-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dc4edaab-1032-4bae-819a-3b74de5bdb00" (UID: "dc4edaab-1032-4bae-819a-3b74de5bdb00"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.755010 4934 scope.go:117] "RemoveContainer" containerID="b5a7f98ab9cc9993b18e3dbf8ba994517fe984ea191805f8373b0da54911f4a0" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.781657 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49vtn\" (UniqueName: \"kubernetes.io/projected/dc4edaab-1032-4bae-819a-3b74de5bdb00-kube-api-access-49vtn\") on node \"crc\" DevicePath \"\"" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.781725 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc4edaab-1032-4bae-819a-3b74de5bdb00-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.781759 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc4edaab-1032-4bae-819a-3b74de5bdb00-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.784880 4934 scope.go:117] "RemoveContainer" containerID="a6d68ed44d2d02956eb58b9cff75e811ad2fe2d892747b73b415eb5e1d72ed16" Oct 02 10:48:01 crc kubenswrapper[4934]: E1002 10:48:01.785499 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6d68ed44d2d02956eb58b9cff75e811ad2fe2d892747b73b415eb5e1d72ed16\": container with ID starting with a6d68ed44d2d02956eb58b9cff75e811ad2fe2d892747b73b415eb5e1d72ed16 not found: ID does not exist" containerID="a6d68ed44d2d02956eb58b9cff75e811ad2fe2d892747b73b415eb5e1d72ed16" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.785570 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6d68ed44d2d02956eb58b9cff75e811ad2fe2d892747b73b415eb5e1d72ed16"} err="failed to get container status \"a6d68ed44d2d02956eb58b9cff75e811ad2fe2d892747b73b415eb5e1d72ed16\": rpc error: code = NotFound desc = could not find container \"a6d68ed44d2d02956eb58b9cff75e811ad2fe2d892747b73b415eb5e1d72ed16\": container with ID starting with a6d68ed44d2d02956eb58b9cff75e811ad2fe2d892747b73b415eb5e1d72ed16 not found: ID does not exist" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.785619 4934 scope.go:117] "RemoveContainer" containerID="7ba52a88e3284066e85ee1b9e1672b6420ccd8abbbd4aa398edd6427d50222dd" Oct 02 10:48:01 crc kubenswrapper[4934]: E1002 10:48:01.786111 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ba52a88e3284066e85ee1b9e1672b6420ccd8abbbd4aa398edd6427d50222dd\": container with ID starting with 7ba52a88e3284066e85ee1b9e1672b6420ccd8abbbd4aa398edd6427d50222dd not found: ID does not exist" containerID="7ba52a88e3284066e85ee1b9e1672b6420ccd8abbbd4aa398edd6427d50222dd" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.786149 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ba52a88e3284066e85ee1b9e1672b6420ccd8abbbd4aa398edd6427d50222dd"} err="failed to get container status \"7ba52a88e3284066e85ee1b9e1672b6420ccd8abbbd4aa398edd6427d50222dd\": rpc error: code = NotFound desc = could not find container \"7ba52a88e3284066e85ee1b9e1672b6420ccd8abbbd4aa398edd6427d50222dd\": container with ID starting with 7ba52a88e3284066e85ee1b9e1672b6420ccd8abbbd4aa398edd6427d50222dd not found: ID does not exist" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.786167 4934 scope.go:117] "RemoveContainer" containerID="b5a7f98ab9cc9993b18e3dbf8ba994517fe984ea191805f8373b0da54911f4a0" Oct 02 10:48:01 crc kubenswrapper[4934]: E1002 10:48:01.786500 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5a7f98ab9cc9993b18e3dbf8ba994517fe984ea191805f8373b0da54911f4a0\": container with ID starting with b5a7f98ab9cc9993b18e3dbf8ba994517fe984ea191805f8373b0da54911f4a0 not found: ID does not exist" containerID="b5a7f98ab9cc9993b18e3dbf8ba994517fe984ea191805f8373b0da54911f4a0" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.786524 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5a7f98ab9cc9993b18e3dbf8ba994517fe984ea191805f8373b0da54911f4a0"} err="failed to get container status \"b5a7f98ab9cc9993b18e3dbf8ba994517fe984ea191805f8373b0da54911f4a0\": rpc error: code = NotFound desc = could not find container \"b5a7f98ab9cc9993b18e3dbf8ba994517fe984ea191805f8373b0da54911f4a0\": container with ID starting with b5a7f98ab9cc9993b18e3dbf8ba994517fe984ea191805f8373b0da54911f4a0 not found: ID does not exist" Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.969789 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wv6td"] Oct 02 10:48:01 crc kubenswrapper[4934]: I1002 10:48:01.974450 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wv6td"] Oct 02 10:48:02 crc kubenswrapper[4934]: I1002 10:48:02.926778 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76c4b5f5-8263-42f8-96e8-e9c730d38330" path="/var/lib/kubelet/pods/76c4b5f5-8263-42f8-96e8-e9c730d38330/volumes" Oct 02 10:48:02 crc kubenswrapper[4934]: I1002 10:48:02.927990 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc4edaab-1032-4bae-819a-3b74de5bdb00" path="/var/lib/kubelet/pods/dc4edaab-1032-4bae-819a-3b74de5bdb00/volumes" Oct 02 10:48:08 crc kubenswrapper[4934]: I1002 10:48:08.439672 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:48:08 crc kubenswrapper[4934]: I1002 10:48:08.440253 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:48:38 crc kubenswrapper[4934]: I1002 10:48:38.439678 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:48:38 crc kubenswrapper[4934]: I1002 10:48:38.440288 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:48:38 crc kubenswrapper[4934]: I1002 10:48:38.440351 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 10:48:38 crc kubenswrapper[4934]: I1002 10:48:38.441229 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d"} pod="openshift-machine-config-operator/machine-config-daemon-djh5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 10:48:38 crc kubenswrapper[4934]: I1002 10:48:38.441353 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" containerID="cri-o://4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d" gracePeriod=600 Oct 02 10:48:38 crc kubenswrapper[4934]: E1002 10:48:38.593326 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:48:38 crc kubenswrapper[4934]: I1002 10:48:38.955928 4934 generic.go:334] "Generic (PLEG): container finished" podID="71db06ef-05b0-4f58-b251-b27117a8500a" containerID="4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d" exitCode=0 Oct 02 10:48:38 crc kubenswrapper[4934]: I1002 10:48:38.955975 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerDied","Data":"4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d"} Oct 02 10:48:38 crc kubenswrapper[4934]: I1002 10:48:38.956011 4934 scope.go:117] "RemoveContainer" containerID="6e65f5d855a69db84d01cc034695cad0d083c86b8bb538609fe0174fad30a308" Oct 02 10:48:38 crc kubenswrapper[4934]: I1002 10:48:38.956639 4934 scope.go:117] "RemoveContainer" containerID="4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d" Oct 02 10:48:38 crc kubenswrapper[4934]: E1002 10:48:38.956859 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:48:53 crc kubenswrapper[4934]: I1002 10:48:53.913188 4934 scope.go:117] "RemoveContainer" containerID="4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d" Oct 02 10:48:53 crc kubenswrapper[4934]: E1002 10:48:53.913772 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:49:06 crc kubenswrapper[4934]: I1002 10:49:06.913469 4934 scope.go:117] "RemoveContainer" containerID="4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d" Oct 02 10:49:06 crc kubenswrapper[4934]: E1002 10:49:06.914302 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:49:19 crc kubenswrapper[4934]: I1002 10:49:19.913059 4934 scope.go:117] "RemoveContainer" containerID="4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d" Oct 02 10:49:19 crc kubenswrapper[4934]: E1002 10:49:19.914128 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:49:32 crc kubenswrapper[4934]: I1002 10:49:32.914312 4934 scope.go:117] "RemoveContainer" containerID="4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d" Oct 02 10:49:32 crc kubenswrapper[4934]: E1002 10:49:32.915332 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:49:43 crc kubenswrapper[4934]: I1002 10:49:43.913286 4934 scope.go:117] "RemoveContainer" containerID="4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d" Oct 02 10:49:43 crc kubenswrapper[4934]: E1002 10:49:43.914313 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:49:57 crc kubenswrapper[4934]: I1002 10:49:57.913827 4934 scope.go:117] "RemoveContainer" containerID="4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d" Oct 02 10:49:57 crc kubenswrapper[4934]: E1002 10:49:57.914940 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:50:08 crc kubenswrapper[4934]: I1002 10:50:08.918502 4934 scope.go:117] "RemoveContainer" containerID="4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d" Oct 02 10:50:08 crc kubenswrapper[4934]: E1002 10:50:08.919257 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:50:20 crc kubenswrapper[4934]: I1002 10:50:20.912754 4934 scope.go:117] "RemoveContainer" containerID="4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d" Oct 02 10:50:20 crc kubenswrapper[4934]: E1002 10:50:20.913604 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:50:34 crc kubenswrapper[4934]: I1002 10:50:34.913470 4934 scope.go:117] "RemoveContainer" containerID="4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d" Oct 02 10:50:34 crc kubenswrapper[4934]: E1002 10:50:34.914120 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:50:45 crc kubenswrapper[4934]: I1002 10:50:45.913367 4934 scope.go:117] "RemoveContainer" containerID="4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d" Oct 02 10:50:45 crc kubenswrapper[4934]: E1002 10:50:45.914528 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:50:57 crc kubenswrapper[4934]: I1002 10:50:57.913403 4934 scope.go:117] "RemoveContainer" containerID="4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d" Oct 02 10:50:57 crc kubenswrapper[4934]: E1002 10:50:57.914150 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:51:10 crc kubenswrapper[4934]: I1002 10:51:10.912892 4934 scope.go:117] "RemoveContainer" containerID="4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d" Oct 02 10:51:10 crc kubenswrapper[4934]: E1002 10:51:10.913735 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:51:24 crc kubenswrapper[4934]: I1002 10:51:24.912984 4934 scope.go:117] "RemoveContainer" containerID="4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d" Oct 02 10:51:24 crc kubenswrapper[4934]: E1002 10:51:24.913589 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:51:39 crc kubenswrapper[4934]: I1002 10:51:39.913540 4934 scope.go:117] "RemoveContainer" containerID="4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d" Oct 02 10:51:39 crc kubenswrapper[4934]: E1002 10:51:39.914464 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:51:53 crc kubenswrapper[4934]: I1002 10:51:53.912939 4934 scope.go:117] "RemoveContainer" containerID="4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d" Oct 02 10:51:53 crc kubenswrapper[4934]: E1002 10:51:53.913548 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:52:04 crc kubenswrapper[4934]: I1002 10:52:04.913178 4934 scope.go:117] "RemoveContainer" containerID="4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d" Oct 02 10:52:04 crc kubenswrapper[4934]: E1002 10:52:04.913840 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:52:08 crc kubenswrapper[4934]: I1002 10:52:08.523729 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b649k"] Oct 02 10:52:08 crc kubenswrapper[4934]: E1002 10:52:08.524494 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc4edaab-1032-4bae-819a-3b74de5bdb00" containerName="registry-server" Oct 02 10:52:08 crc kubenswrapper[4934]: I1002 10:52:08.524512 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc4edaab-1032-4bae-819a-3b74de5bdb00" containerName="registry-server" Oct 02 10:52:08 crc kubenswrapper[4934]: E1002 10:52:08.524524 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc4edaab-1032-4bae-819a-3b74de5bdb00" containerName="extract-utilities" Oct 02 10:52:08 crc kubenswrapper[4934]: I1002 10:52:08.524535 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc4edaab-1032-4bae-819a-3b74de5bdb00" containerName="extract-utilities" Oct 02 10:52:08 crc kubenswrapper[4934]: E1002 10:52:08.524551 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76c4b5f5-8263-42f8-96e8-e9c730d38330" containerName="extract-utilities" Oct 02 10:52:08 crc kubenswrapper[4934]: I1002 10:52:08.524561 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="76c4b5f5-8263-42f8-96e8-e9c730d38330" containerName="extract-utilities" Oct 02 10:52:08 crc kubenswrapper[4934]: E1002 10:52:08.524604 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76c4b5f5-8263-42f8-96e8-e9c730d38330" containerName="extract-content" Oct 02 10:52:08 crc kubenswrapper[4934]: I1002 10:52:08.524614 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="76c4b5f5-8263-42f8-96e8-e9c730d38330" containerName="extract-content" Oct 02 10:52:08 crc kubenswrapper[4934]: E1002 10:52:08.524634 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc4edaab-1032-4bae-819a-3b74de5bdb00" containerName="extract-content" Oct 02 10:52:08 crc kubenswrapper[4934]: I1002 10:52:08.524643 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc4edaab-1032-4bae-819a-3b74de5bdb00" containerName="extract-content" Oct 02 10:52:08 crc kubenswrapper[4934]: E1002 10:52:08.524673 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76c4b5f5-8263-42f8-96e8-e9c730d38330" containerName="registry-server" Oct 02 10:52:08 crc kubenswrapper[4934]: I1002 10:52:08.524683 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="76c4b5f5-8263-42f8-96e8-e9c730d38330" containerName="registry-server" Oct 02 10:52:08 crc kubenswrapper[4934]: I1002 10:52:08.524904 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="76c4b5f5-8263-42f8-96e8-e9c730d38330" containerName="registry-server" Oct 02 10:52:08 crc kubenswrapper[4934]: I1002 10:52:08.524939 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc4edaab-1032-4bae-819a-3b74de5bdb00" containerName="registry-server" Oct 02 10:52:08 crc kubenswrapper[4934]: I1002 10:52:08.527952 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b649k" Oct 02 10:52:08 crc kubenswrapper[4934]: I1002 10:52:08.546320 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b649k"] Oct 02 10:52:08 crc kubenswrapper[4934]: I1002 10:52:08.641875 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bac4fc89-9507-428d-b366-139805239f07-utilities\") pod \"redhat-marketplace-b649k\" (UID: \"bac4fc89-9507-428d-b366-139805239f07\") " pod="openshift-marketplace/redhat-marketplace-b649k" Oct 02 10:52:08 crc kubenswrapper[4934]: I1002 10:52:08.641961 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s4ft\" (UniqueName: \"kubernetes.io/projected/bac4fc89-9507-428d-b366-139805239f07-kube-api-access-2s4ft\") pod \"redhat-marketplace-b649k\" (UID: \"bac4fc89-9507-428d-b366-139805239f07\") " pod="openshift-marketplace/redhat-marketplace-b649k" Oct 02 10:52:08 crc kubenswrapper[4934]: I1002 10:52:08.641991 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bac4fc89-9507-428d-b366-139805239f07-catalog-content\") pod \"redhat-marketplace-b649k\" (UID: \"bac4fc89-9507-428d-b366-139805239f07\") " pod="openshift-marketplace/redhat-marketplace-b649k" Oct 02 10:52:08 crc kubenswrapper[4934]: I1002 10:52:08.743336 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s4ft\" (UniqueName: \"kubernetes.io/projected/bac4fc89-9507-428d-b366-139805239f07-kube-api-access-2s4ft\") pod \"redhat-marketplace-b649k\" (UID: \"bac4fc89-9507-428d-b366-139805239f07\") " pod="openshift-marketplace/redhat-marketplace-b649k" Oct 02 10:52:08 crc kubenswrapper[4934]: I1002 10:52:08.743447 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bac4fc89-9507-428d-b366-139805239f07-catalog-content\") pod \"redhat-marketplace-b649k\" (UID: \"bac4fc89-9507-428d-b366-139805239f07\") " pod="openshift-marketplace/redhat-marketplace-b649k" Oct 02 10:52:08 crc kubenswrapper[4934]: I1002 10:52:08.743681 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bac4fc89-9507-428d-b366-139805239f07-utilities\") pod \"redhat-marketplace-b649k\" (UID: \"bac4fc89-9507-428d-b366-139805239f07\") " pod="openshift-marketplace/redhat-marketplace-b649k" Oct 02 10:52:08 crc kubenswrapper[4934]: I1002 10:52:08.744036 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bac4fc89-9507-428d-b366-139805239f07-catalog-content\") pod \"redhat-marketplace-b649k\" (UID: \"bac4fc89-9507-428d-b366-139805239f07\") " pod="openshift-marketplace/redhat-marketplace-b649k" Oct 02 10:52:08 crc kubenswrapper[4934]: I1002 10:52:08.744060 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bac4fc89-9507-428d-b366-139805239f07-utilities\") pod \"redhat-marketplace-b649k\" (UID: \"bac4fc89-9507-428d-b366-139805239f07\") " pod="openshift-marketplace/redhat-marketplace-b649k" Oct 02 10:52:08 crc kubenswrapper[4934]: I1002 10:52:08.777935 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s4ft\" (UniqueName: \"kubernetes.io/projected/bac4fc89-9507-428d-b366-139805239f07-kube-api-access-2s4ft\") pod \"redhat-marketplace-b649k\" (UID: \"bac4fc89-9507-428d-b366-139805239f07\") " pod="openshift-marketplace/redhat-marketplace-b649k" Oct 02 10:52:08 crc kubenswrapper[4934]: I1002 10:52:08.860791 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b649k" Oct 02 10:52:09 crc kubenswrapper[4934]: I1002 10:52:09.321781 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b649k"] Oct 02 10:52:09 crc kubenswrapper[4934]: I1002 10:52:09.778992 4934 generic.go:334] "Generic (PLEG): container finished" podID="bac4fc89-9507-428d-b366-139805239f07" containerID="5187d7d3c0caff33cbb66a037292e879243cbdaf8800e2e52d22295a1cd366d6" exitCode=0 Oct 02 10:52:09 crc kubenswrapper[4934]: I1002 10:52:09.779089 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b649k" event={"ID":"bac4fc89-9507-428d-b366-139805239f07","Type":"ContainerDied","Data":"5187d7d3c0caff33cbb66a037292e879243cbdaf8800e2e52d22295a1cd366d6"} Oct 02 10:52:09 crc kubenswrapper[4934]: I1002 10:52:09.779296 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b649k" event={"ID":"bac4fc89-9507-428d-b366-139805239f07","Type":"ContainerStarted","Data":"451760f32bd34e7f06b134b8ca7c9ed8fac7dab8160d7ecf2304c0f2efaa7898"} Oct 02 10:52:09 crc kubenswrapper[4934]: I1002 10:52:09.781354 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 10:52:10 crc kubenswrapper[4934]: I1002 10:52:10.788936 4934 generic.go:334] "Generic (PLEG): container finished" podID="bac4fc89-9507-428d-b366-139805239f07" containerID="5d71b4ead56b66bc9c96918e2d4acfd2895ea6e2746d73090f51a17d1f7c85f8" exitCode=0 Oct 02 10:52:10 crc kubenswrapper[4934]: I1002 10:52:10.789235 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b649k" event={"ID":"bac4fc89-9507-428d-b366-139805239f07","Type":"ContainerDied","Data":"5d71b4ead56b66bc9c96918e2d4acfd2895ea6e2746d73090f51a17d1f7c85f8"} Oct 02 10:52:11 crc kubenswrapper[4934]: I1002 10:52:11.799783 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b649k" event={"ID":"bac4fc89-9507-428d-b366-139805239f07","Type":"ContainerStarted","Data":"5e0d68e1a63dc2bf5105ca7412ce21574947227d5c7baac7c7423595a214bdc7"} Oct 02 10:52:11 crc kubenswrapper[4934]: I1002 10:52:11.829731 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b649k" podStartSLOduration=2.414407723 podStartE2EDuration="3.82971312s" podCreationTimestamp="2025-10-02 10:52:08 +0000 UTC" firstStartedPulling="2025-10-02 10:52:09.780993708 +0000 UTC m=+3801.533635270" lastFinishedPulling="2025-10-02 10:52:11.196299105 +0000 UTC m=+3802.948940667" observedRunningTime="2025-10-02 10:52:11.824763984 +0000 UTC m=+3803.577405526" watchObservedRunningTime="2025-10-02 10:52:11.82971312 +0000 UTC m=+3803.582354652" Oct 02 10:52:15 crc kubenswrapper[4934]: I1002 10:52:15.912757 4934 scope.go:117] "RemoveContainer" containerID="4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d" Oct 02 10:52:15 crc kubenswrapper[4934]: E1002 10:52:15.913705 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:52:18 crc kubenswrapper[4934]: I1002 10:52:18.861150 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b649k" Oct 02 10:52:18 crc kubenswrapper[4934]: I1002 10:52:18.861534 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b649k" Oct 02 10:52:18 crc kubenswrapper[4934]: I1002 10:52:18.921971 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b649k" Oct 02 10:52:19 crc kubenswrapper[4934]: I1002 10:52:19.927914 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b649k" Oct 02 10:52:19 crc kubenswrapper[4934]: I1002 10:52:19.998147 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b649k"] Oct 02 10:52:21 crc kubenswrapper[4934]: I1002 10:52:21.898644 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b649k" podUID="bac4fc89-9507-428d-b366-139805239f07" containerName="registry-server" containerID="cri-o://5e0d68e1a63dc2bf5105ca7412ce21574947227d5c7baac7c7423595a214bdc7" gracePeriod=2 Oct 02 10:52:22 crc kubenswrapper[4934]: I1002 10:52:22.289650 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b649k" Oct 02 10:52:22 crc kubenswrapper[4934]: I1002 10:52:22.445608 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bac4fc89-9507-428d-b366-139805239f07-catalog-content\") pod \"bac4fc89-9507-428d-b366-139805239f07\" (UID: \"bac4fc89-9507-428d-b366-139805239f07\") " Oct 02 10:52:22 crc kubenswrapper[4934]: I1002 10:52:22.445676 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bac4fc89-9507-428d-b366-139805239f07-utilities\") pod \"bac4fc89-9507-428d-b366-139805239f07\" (UID: \"bac4fc89-9507-428d-b366-139805239f07\") " Oct 02 10:52:22 crc kubenswrapper[4934]: I1002 10:52:22.445717 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2s4ft\" (UniqueName: \"kubernetes.io/projected/bac4fc89-9507-428d-b366-139805239f07-kube-api-access-2s4ft\") pod \"bac4fc89-9507-428d-b366-139805239f07\" (UID: \"bac4fc89-9507-428d-b366-139805239f07\") " Oct 02 10:52:22 crc kubenswrapper[4934]: I1002 10:52:22.446850 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bac4fc89-9507-428d-b366-139805239f07-utilities" (OuterVolumeSpecName: "utilities") pod "bac4fc89-9507-428d-b366-139805239f07" (UID: "bac4fc89-9507-428d-b366-139805239f07"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:52:22 crc kubenswrapper[4934]: I1002 10:52:22.453919 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bac4fc89-9507-428d-b366-139805239f07-kube-api-access-2s4ft" (OuterVolumeSpecName: "kube-api-access-2s4ft") pod "bac4fc89-9507-428d-b366-139805239f07" (UID: "bac4fc89-9507-428d-b366-139805239f07"). InnerVolumeSpecName "kube-api-access-2s4ft". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:52:22 crc kubenswrapper[4934]: I1002 10:52:22.472991 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bac4fc89-9507-428d-b366-139805239f07-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bac4fc89-9507-428d-b366-139805239f07" (UID: "bac4fc89-9507-428d-b366-139805239f07"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:52:22 crc kubenswrapper[4934]: I1002 10:52:22.546895 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bac4fc89-9507-428d-b366-139805239f07-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:22 crc kubenswrapper[4934]: I1002 10:52:22.546930 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bac4fc89-9507-428d-b366-139805239f07-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:22 crc kubenswrapper[4934]: I1002 10:52:22.546940 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2s4ft\" (UniqueName: \"kubernetes.io/projected/bac4fc89-9507-428d-b366-139805239f07-kube-api-access-2s4ft\") on node \"crc\" DevicePath \"\"" Oct 02 10:52:22 crc kubenswrapper[4934]: I1002 10:52:22.912353 4934 generic.go:334] "Generic (PLEG): container finished" podID="bac4fc89-9507-428d-b366-139805239f07" containerID="5e0d68e1a63dc2bf5105ca7412ce21574947227d5c7baac7c7423595a214bdc7" exitCode=0 Oct 02 10:52:22 crc kubenswrapper[4934]: I1002 10:52:22.912428 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b649k" Oct 02 10:52:22 crc kubenswrapper[4934]: I1002 10:52:22.925414 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b649k" event={"ID":"bac4fc89-9507-428d-b366-139805239f07","Type":"ContainerDied","Data":"5e0d68e1a63dc2bf5105ca7412ce21574947227d5c7baac7c7423595a214bdc7"} Oct 02 10:52:22 crc kubenswrapper[4934]: I1002 10:52:22.925457 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b649k" event={"ID":"bac4fc89-9507-428d-b366-139805239f07","Type":"ContainerDied","Data":"451760f32bd34e7f06b134b8ca7c9ed8fac7dab8160d7ecf2304c0f2efaa7898"} Oct 02 10:52:22 crc kubenswrapper[4934]: I1002 10:52:22.925477 4934 scope.go:117] "RemoveContainer" containerID="5e0d68e1a63dc2bf5105ca7412ce21574947227d5c7baac7c7423595a214bdc7" Oct 02 10:52:22 crc kubenswrapper[4934]: I1002 10:52:22.967104 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b649k"] Oct 02 10:52:22 crc kubenswrapper[4934]: I1002 10:52:22.967423 4934 scope.go:117] "RemoveContainer" containerID="5d71b4ead56b66bc9c96918e2d4acfd2895ea6e2746d73090f51a17d1f7c85f8" Oct 02 10:52:22 crc kubenswrapper[4934]: I1002 10:52:22.974034 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b649k"] Oct 02 10:52:22 crc kubenswrapper[4934]: I1002 10:52:22.998862 4934 scope.go:117] "RemoveContainer" containerID="5187d7d3c0caff33cbb66a037292e879243cbdaf8800e2e52d22295a1cd366d6" Oct 02 10:52:23 crc kubenswrapper[4934]: I1002 10:52:23.037145 4934 scope.go:117] "RemoveContainer" containerID="5e0d68e1a63dc2bf5105ca7412ce21574947227d5c7baac7c7423595a214bdc7" Oct 02 10:52:23 crc kubenswrapper[4934]: E1002 10:52:23.038096 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e0d68e1a63dc2bf5105ca7412ce21574947227d5c7baac7c7423595a214bdc7\": container with ID starting with 5e0d68e1a63dc2bf5105ca7412ce21574947227d5c7baac7c7423595a214bdc7 not found: ID does not exist" containerID="5e0d68e1a63dc2bf5105ca7412ce21574947227d5c7baac7c7423595a214bdc7" Oct 02 10:52:23 crc kubenswrapper[4934]: I1002 10:52:23.038245 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e0d68e1a63dc2bf5105ca7412ce21574947227d5c7baac7c7423595a214bdc7"} err="failed to get container status \"5e0d68e1a63dc2bf5105ca7412ce21574947227d5c7baac7c7423595a214bdc7\": rpc error: code = NotFound desc = could not find container \"5e0d68e1a63dc2bf5105ca7412ce21574947227d5c7baac7c7423595a214bdc7\": container with ID starting with 5e0d68e1a63dc2bf5105ca7412ce21574947227d5c7baac7c7423595a214bdc7 not found: ID does not exist" Oct 02 10:52:23 crc kubenswrapper[4934]: I1002 10:52:23.038400 4934 scope.go:117] "RemoveContainer" containerID="5d71b4ead56b66bc9c96918e2d4acfd2895ea6e2746d73090f51a17d1f7c85f8" Oct 02 10:52:23 crc kubenswrapper[4934]: E1002 10:52:23.038954 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d71b4ead56b66bc9c96918e2d4acfd2895ea6e2746d73090f51a17d1f7c85f8\": container with ID starting with 5d71b4ead56b66bc9c96918e2d4acfd2895ea6e2746d73090f51a17d1f7c85f8 not found: ID does not exist" containerID="5d71b4ead56b66bc9c96918e2d4acfd2895ea6e2746d73090f51a17d1f7c85f8" Oct 02 10:52:23 crc kubenswrapper[4934]: I1002 10:52:23.039061 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d71b4ead56b66bc9c96918e2d4acfd2895ea6e2746d73090f51a17d1f7c85f8"} err="failed to get container status \"5d71b4ead56b66bc9c96918e2d4acfd2895ea6e2746d73090f51a17d1f7c85f8\": rpc error: code = NotFound desc = could not find container \"5d71b4ead56b66bc9c96918e2d4acfd2895ea6e2746d73090f51a17d1f7c85f8\": container with ID starting with 5d71b4ead56b66bc9c96918e2d4acfd2895ea6e2746d73090f51a17d1f7c85f8 not found: ID does not exist" Oct 02 10:52:23 crc kubenswrapper[4934]: I1002 10:52:23.039147 4934 scope.go:117] "RemoveContainer" containerID="5187d7d3c0caff33cbb66a037292e879243cbdaf8800e2e52d22295a1cd366d6" Oct 02 10:52:23 crc kubenswrapper[4934]: E1002 10:52:23.039487 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5187d7d3c0caff33cbb66a037292e879243cbdaf8800e2e52d22295a1cd366d6\": container with ID starting with 5187d7d3c0caff33cbb66a037292e879243cbdaf8800e2e52d22295a1cd366d6 not found: ID does not exist" containerID="5187d7d3c0caff33cbb66a037292e879243cbdaf8800e2e52d22295a1cd366d6" Oct 02 10:52:23 crc kubenswrapper[4934]: I1002 10:52:23.039602 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5187d7d3c0caff33cbb66a037292e879243cbdaf8800e2e52d22295a1cd366d6"} err="failed to get container status \"5187d7d3c0caff33cbb66a037292e879243cbdaf8800e2e52d22295a1cd366d6\": rpc error: code = NotFound desc = could not find container \"5187d7d3c0caff33cbb66a037292e879243cbdaf8800e2e52d22295a1cd366d6\": container with ID starting with 5187d7d3c0caff33cbb66a037292e879243cbdaf8800e2e52d22295a1cd366d6 not found: ID does not exist" Oct 02 10:52:24 crc kubenswrapper[4934]: I1002 10:52:24.930301 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bac4fc89-9507-428d-b366-139805239f07" path="/var/lib/kubelet/pods/bac4fc89-9507-428d-b366-139805239f07/volumes" Oct 02 10:52:29 crc kubenswrapper[4934]: I1002 10:52:29.913006 4934 scope.go:117] "RemoveContainer" containerID="4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d" Oct 02 10:52:29 crc kubenswrapper[4934]: E1002 10:52:29.913657 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:52:43 crc kubenswrapper[4934]: I1002 10:52:43.912745 4934 scope.go:117] "RemoveContainer" containerID="4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d" Oct 02 10:52:43 crc kubenswrapper[4934]: E1002 10:52:43.913876 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:52:58 crc kubenswrapper[4934]: I1002 10:52:58.917639 4934 scope.go:117] "RemoveContainer" containerID="4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d" Oct 02 10:52:58 crc kubenswrapper[4934]: E1002 10:52:58.918340 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:53:08 crc kubenswrapper[4934]: I1002 10:53:08.132520 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9hdgv"] Oct 02 10:53:08 crc kubenswrapper[4934]: E1002 10:53:08.133531 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bac4fc89-9507-428d-b366-139805239f07" containerName="extract-utilities" Oct 02 10:53:08 crc kubenswrapper[4934]: I1002 10:53:08.133554 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="bac4fc89-9507-428d-b366-139805239f07" containerName="extract-utilities" Oct 02 10:53:08 crc kubenswrapper[4934]: E1002 10:53:08.133607 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bac4fc89-9507-428d-b366-139805239f07" containerName="registry-server" Oct 02 10:53:08 crc kubenswrapper[4934]: I1002 10:53:08.133620 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="bac4fc89-9507-428d-b366-139805239f07" containerName="registry-server" Oct 02 10:53:08 crc kubenswrapper[4934]: E1002 10:53:08.133636 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bac4fc89-9507-428d-b366-139805239f07" containerName="extract-content" Oct 02 10:53:08 crc kubenswrapper[4934]: I1002 10:53:08.133648 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="bac4fc89-9507-428d-b366-139805239f07" containerName="extract-content" Oct 02 10:53:08 crc kubenswrapper[4934]: I1002 10:53:08.133864 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="bac4fc89-9507-428d-b366-139805239f07" containerName="registry-server" Oct 02 10:53:08 crc kubenswrapper[4934]: I1002 10:53:08.135146 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9hdgv" Oct 02 10:53:08 crc kubenswrapper[4934]: I1002 10:53:08.149536 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9hdgv"] Oct 02 10:53:08 crc kubenswrapper[4934]: I1002 10:53:08.258854 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfg7q\" (UniqueName: \"kubernetes.io/projected/d1826925-c5fc-4e16-a51b-da8e36c6ffcd-kube-api-access-jfg7q\") pod \"community-operators-9hdgv\" (UID: \"d1826925-c5fc-4e16-a51b-da8e36c6ffcd\") " pod="openshift-marketplace/community-operators-9hdgv" Oct 02 10:53:08 crc kubenswrapper[4934]: I1002 10:53:08.258928 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1826925-c5fc-4e16-a51b-da8e36c6ffcd-utilities\") pod \"community-operators-9hdgv\" (UID: \"d1826925-c5fc-4e16-a51b-da8e36c6ffcd\") " pod="openshift-marketplace/community-operators-9hdgv" Oct 02 10:53:08 crc kubenswrapper[4934]: I1002 10:53:08.259009 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1826925-c5fc-4e16-a51b-da8e36c6ffcd-catalog-content\") pod \"community-operators-9hdgv\" (UID: \"d1826925-c5fc-4e16-a51b-da8e36c6ffcd\") " pod="openshift-marketplace/community-operators-9hdgv" Oct 02 10:53:08 crc kubenswrapper[4934]: I1002 10:53:08.360218 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1826925-c5fc-4e16-a51b-da8e36c6ffcd-catalog-content\") pod \"community-operators-9hdgv\" (UID: \"d1826925-c5fc-4e16-a51b-da8e36c6ffcd\") " pod="openshift-marketplace/community-operators-9hdgv" Oct 02 10:53:08 crc kubenswrapper[4934]: I1002 10:53:08.360270 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfg7q\" (UniqueName: \"kubernetes.io/projected/d1826925-c5fc-4e16-a51b-da8e36c6ffcd-kube-api-access-jfg7q\") pod \"community-operators-9hdgv\" (UID: \"d1826925-c5fc-4e16-a51b-da8e36c6ffcd\") " pod="openshift-marketplace/community-operators-9hdgv" Oct 02 10:53:08 crc kubenswrapper[4934]: I1002 10:53:08.360309 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1826925-c5fc-4e16-a51b-da8e36c6ffcd-utilities\") pod \"community-operators-9hdgv\" (UID: \"d1826925-c5fc-4e16-a51b-da8e36c6ffcd\") " pod="openshift-marketplace/community-operators-9hdgv" Oct 02 10:53:08 crc kubenswrapper[4934]: I1002 10:53:08.360868 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1826925-c5fc-4e16-a51b-da8e36c6ffcd-utilities\") pod \"community-operators-9hdgv\" (UID: \"d1826925-c5fc-4e16-a51b-da8e36c6ffcd\") " pod="openshift-marketplace/community-operators-9hdgv" Oct 02 10:53:08 crc kubenswrapper[4934]: I1002 10:53:08.361051 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1826925-c5fc-4e16-a51b-da8e36c6ffcd-catalog-content\") pod \"community-operators-9hdgv\" (UID: \"d1826925-c5fc-4e16-a51b-da8e36c6ffcd\") " pod="openshift-marketplace/community-operators-9hdgv" Oct 02 10:53:08 crc kubenswrapper[4934]: I1002 10:53:08.385416 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfg7q\" (UniqueName: \"kubernetes.io/projected/d1826925-c5fc-4e16-a51b-da8e36c6ffcd-kube-api-access-jfg7q\") pod \"community-operators-9hdgv\" (UID: \"d1826925-c5fc-4e16-a51b-da8e36c6ffcd\") " pod="openshift-marketplace/community-operators-9hdgv" Oct 02 10:53:08 crc kubenswrapper[4934]: I1002 10:53:08.461429 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9hdgv" Oct 02 10:53:08 crc kubenswrapper[4934]: I1002 10:53:08.945611 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9hdgv"] Oct 02 10:53:09 crc kubenswrapper[4934]: W1002 10:53:09.020072 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1826925_c5fc_4e16_a51b_da8e36c6ffcd.slice/crio-671f14a28ec166ce058ae9d40a69e3a1f58548b321b56dc9889b126e2596cc38 WatchSource:0}: Error finding container 671f14a28ec166ce058ae9d40a69e3a1f58548b321b56dc9889b126e2596cc38: Status 404 returned error can't find the container with id 671f14a28ec166ce058ae9d40a69e3a1f58548b321b56dc9889b126e2596cc38 Oct 02 10:53:09 crc kubenswrapper[4934]: I1002 10:53:09.293911 4934 generic.go:334] "Generic (PLEG): container finished" podID="d1826925-c5fc-4e16-a51b-da8e36c6ffcd" containerID="dcdbb7a09d7323ed1d9f664606b00099ab84379ee4e768568d43530fbd607ab9" exitCode=0 Oct 02 10:53:09 crc kubenswrapper[4934]: I1002 10:53:09.294019 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9hdgv" event={"ID":"d1826925-c5fc-4e16-a51b-da8e36c6ffcd","Type":"ContainerDied","Data":"dcdbb7a09d7323ed1d9f664606b00099ab84379ee4e768568d43530fbd607ab9"} Oct 02 10:53:09 crc kubenswrapper[4934]: I1002 10:53:09.294186 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9hdgv" event={"ID":"d1826925-c5fc-4e16-a51b-da8e36c6ffcd","Type":"ContainerStarted","Data":"671f14a28ec166ce058ae9d40a69e3a1f58548b321b56dc9889b126e2596cc38"} Oct 02 10:53:12 crc kubenswrapper[4934]: I1002 10:53:12.321802 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9hdgv" event={"ID":"d1826925-c5fc-4e16-a51b-da8e36c6ffcd","Type":"ContainerStarted","Data":"7dcb97ee85f738e224a2b80adcf8cc9238632bdc5b8e2ebb242427461681b486"} Oct 02 10:53:12 crc kubenswrapper[4934]: I1002 10:53:12.913627 4934 scope.go:117] "RemoveContainer" containerID="4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d" Oct 02 10:53:12 crc kubenswrapper[4934]: E1002 10:53:12.914193 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:53:13 crc kubenswrapper[4934]: I1002 10:53:13.334019 4934 generic.go:334] "Generic (PLEG): container finished" podID="d1826925-c5fc-4e16-a51b-da8e36c6ffcd" containerID="7dcb97ee85f738e224a2b80adcf8cc9238632bdc5b8e2ebb242427461681b486" exitCode=0 Oct 02 10:53:13 crc kubenswrapper[4934]: I1002 10:53:13.334863 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9hdgv" event={"ID":"d1826925-c5fc-4e16-a51b-da8e36c6ffcd","Type":"ContainerDied","Data":"7dcb97ee85f738e224a2b80adcf8cc9238632bdc5b8e2ebb242427461681b486"} Oct 02 10:53:16 crc kubenswrapper[4934]: I1002 10:53:16.363432 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9hdgv" event={"ID":"d1826925-c5fc-4e16-a51b-da8e36c6ffcd","Type":"ContainerStarted","Data":"d3a6fe0cb3b008bfe59c918adaf4b941558112628e74bbf977144b34d72190a3"} Oct 02 10:53:16 crc kubenswrapper[4934]: I1002 10:53:16.389888 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9hdgv" podStartSLOduration=2.8466061639999998 podStartE2EDuration="8.389868214s" podCreationTimestamp="2025-10-02 10:53:08 +0000 UTC" firstStartedPulling="2025-10-02 10:53:09.295522546 +0000 UTC m=+3861.048164068" lastFinishedPulling="2025-10-02 10:53:14.838784596 +0000 UTC m=+3866.591426118" observedRunningTime="2025-10-02 10:53:16.386335307 +0000 UTC m=+3868.138976839" watchObservedRunningTime="2025-10-02 10:53:16.389868214 +0000 UTC m=+3868.142509736" Oct 02 10:53:18 crc kubenswrapper[4934]: I1002 10:53:18.461837 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9hdgv" Oct 02 10:53:18 crc kubenswrapper[4934]: I1002 10:53:18.461892 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9hdgv" Oct 02 10:53:18 crc kubenswrapper[4934]: I1002 10:53:18.510834 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9hdgv" Oct 02 10:53:26 crc kubenswrapper[4934]: I1002 10:53:26.913702 4934 scope.go:117] "RemoveContainer" containerID="4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d" Oct 02 10:53:26 crc kubenswrapper[4934]: E1002 10:53:26.914691 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 10:53:28 crc kubenswrapper[4934]: I1002 10:53:28.541956 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9hdgv" Oct 02 10:53:28 crc kubenswrapper[4934]: I1002 10:53:28.593397 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9hdgv"] Oct 02 10:53:29 crc kubenswrapper[4934]: I1002 10:53:29.464360 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9hdgv" podUID="d1826925-c5fc-4e16-a51b-da8e36c6ffcd" containerName="registry-server" containerID="cri-o://d3a6fe0cb3b008bfe59c918adaf4b941558112628e74bbf977144b34d72190a3" gracePeriod=2 Oct 02 10:53:30 crc kubenswrapper[4934]: I1002 10:53:30.473618 4934 generic.go:334] "Generic (PLEG): container finished" podID="d1826925-c5fc-4e16-a51b-da8e36c6ffcd" containerID="d3a6fe0cb3b008bfe59c918adaf4b941558112628e74bbf977144b34d72190a3" exitCode=0 Oct 02 10:53:30 crc kubenswrapper[4934]: I1002 10:53:30.473657 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9hdgv" event={"ID":"d1826925-c5fc-4e16-a51b-da8e36c6ffcd","Type":"ContainerDied","Data":"d3a6fe0cb3b008bfe59c918adaf4b941558112628e74bbf977144b34d72190a3"} Oct 02 10:53:30 crc kubenswrapper[4934]: I1002 10:53:30.473989 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9hdgv" event={"ID":"d1826925-c5fc-4e16-a51b-da8e36c6ffcd","Type":"ContainerDied","Data":"671f14a28ec166ce058ae9d40a69e3a1f58548b321b56dc9889b126e2596cc38"} Oct 02 10:53:30 crc kubenswrapper[4934]: I1002 10:53:30.474009 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="671f14a28ec166ce058ae9d40a69e3a1f58548b321b56dc9889b126e2596cc38" Oct 02 10:53:30 crc kubenswrapper[4934]: I1002 10:53:30.494313 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9hdgv" Oct 02 10:53:30 crc kubenswrapper[4934]: I1002 10:53:30.514039 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfg7q\" (UniqueName: \"kubernetes.io/projected/d1826925-c5fc-4e16-a51b-da8e36c6ffcd-kube-api-access-jfg7q\") pod \"d1826925-c5fc-4e16-a51b-da8e36c6ffcd\" (UID: \"d1826925-c5fc-4e16-a51b-da8e36c6ffcd\") " Oct 02 10:53:30 crc kubenswrapper[4934]: I1002 10:53:30.514118 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1826925-c5fc-4e16-a51b-da8e36c6ffcd-utilities\") pod \"d1826925-c5fc-4e16-a51b-da8e36c6ffcd\" (UID: \"d1826925-c5fc-4e16-a51b-da8e36c6ffcd\") " Oct 02 10:53:30 crc kubenswrapper[4934]: I1002 10:53:30.514158 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1826925-c5fc-4e16-a51b-da8e36c6ffcd-catalog-content\") pod \"d1826925-c5fc-4e16-a51b-da8e36c6ffcd\" (UID: \"d1826925-c5fc-4e16-a51b-da8e36c6ffcd\") " Oct 02 10:53:30 crc kubenswrapper[4934]: I1002 10:53:30.561114 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1826925-c5fc-4e16-a51b-da8e36c6ffcd-utilities" (OuterVolumeSpecName: "utilities") pod "d1826925-c5fc-4e16-a51b-da8e36c6ffcd" (UID: "d1826925-c5fc-4e16-a51b-da8e36c6ffcd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:53:30 crc kubenswrapper[4934]: I1002 10:53:30.561634 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1826925-c5fc-4e16-a51b-da8e36c6ffcd-kube-api-access-jfg7q" (OuterVolumeSpecName: "kube-api-access-jfg7q") pod "d1826925-c5fc-4e16-a51b-da8e36c6ffcd" (UID: "d1826925-c5fc-4e16-a51b-da8e36c6ffcd"). InnerVolumeSpecName "kube-api-access-jfg7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:53:30 crc kubenswrapper[4934]: I1002 10:53:30.573964 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1826925-c5fc-4e16-a51b-da8e36c6ffcd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d1826925-c5fc-4e16-a51b-da8e36c6ffcd" (UID: "d1826925-c5fc-4e16-a51b-da8e36c6ffcd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:53:30 crc kubenswrapper[4934]: I1002 10:53:30.615384 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1826925-c5fc-4e16-a51b-da8e36c6ffcd-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:53:30 crc kubenswrapper[4934]: I1002 10:53:30.615423 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1826925-c5fc-4e16-a51b-da8e36c6ffcd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:53:30 crc kubenswrapper[4934]: I1002 10:53:30.615438 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfg7q\" (UniqueName: \"kubernetes.io/projected/d1826925-c5fc-4e16-a51b-da8e36c6ffcd-kube-api-access-jfg7q\") on node \"crc\" DevicePath \"\"" Oct 02 10:53:31 crc kubenswrapper[4934]: I1002 10:53:31.483995 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9hdgv" Oct 02 10:53:31 crc kubenswrapper[4934]: I1002 10:53:31.509430 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9hdgv"] Oct 02 10:53:31 crc kubenswrapper[4934]: I1002 10:53:31.515862 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9hdgv"] Oct 02 10:53:32 crc kubenswrapper[4934]: I1002 10:53:32.939069 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1826925-c5fc-4e16-a51b-da8e36c6ffcd" path="/var/lib/kubelet/pods/d1826925-c5fc-4e16-a51b-da8e36c6ffcd/volumes" Oct 02 10:53:40 crc kubenswrapper[4934]: I1002 10:53:40.913640 4934 scope.go:117] "RemoveContainer" containerID="4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d" Oct 02 10:53:41 crc kubenswrapper[4934]: I1002 10:53:41.584770 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerStarted","Data":"acd6fc421b82d19f0da3591bd2215c6dd60b927b95d1597f6b8f779ce91c0f24"} Oct 02 10:56:08 crc kubenswrapper[4934]: I1002 10:56:08.439611 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:56:08 crc kubenswrapper[4934]: I1002 10:56:08.440356 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:56:38 crc kubenswrapper[4934]: I1002 10:56:38.439719 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:56:38 crc kubenswrapper[4934]: I1002 10:56:38.440234 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:57:08 crc kubenswrapper[4934]: I1002 10:57:08.439705 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:57:08 crc kubenswrapper[4934]: I1002 10:57:08.440417 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:57:08 crc kubenswrapper[4934]: I1002 10:57:08.440483 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 10:57:08 crc kubenswrapper[4934]: I1002 10:57:08.441280 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"acd6fc421b82d19f0da3591bd2215c6dd60b927b95d1597f6b8f779ce91c0f24"} pod="openshift-machine-config-operator/machine-config-daemon-djh5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 10:57:08 crc kubenswrapper[4934]: I1002 10:57:08.441351 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" containerID="cri-o://acd6fc421b82d19f0da3591bd2215c6dd60b927b95d1597f6b8f779ce91c0f24" gracePeriod=600 Oct 02 10:57:09 crc kubenswrapper[4934]: I1002 10:57:09.369759 4934 generic.go:334] "Generic (PLEG): container finished" podID="71db06ef-05b0-4f58-b251-b27117a8500a" containerID="acd6fc421b82d19f0da3591bd2215c6dd60b927b95d1597f6b8f779ce91c0f24" exitCode=0 Oct 02 10:57:09 crc kubenswrapper[4934]: I1002 10:57:09.369955 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerDied","Data":"acd6fc421b82d19f0da3591bd2215c6dd60b927b95d1597f6b8f779ce91c0f24"} Oct 02 10:57:09 crc kubenswrapper[4934]: I1002 10:57:09.370203 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerStarted","Data":"ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39"} Oct 02 10:57:09 crc kubenswrapper[4934]: I1002 10:57:09.370233 4934 scope.go:117] "RemoveContainer" containerID="4e23930340580f338468c54fbccad0c764a053b063ffe5f8fe76ad42df45783d" Oct 02 10:58:20 crc kubenswrapper[4934]: I1002 10:58:20.860569 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7677s"] Oct 02 10:58:20 crc kubenswrapper[4934]: E1002 10:58:20.862769 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1826925-c5fc-4e16-a51b-da8e36c6ffcd" containerName="extract-content" Oct 02 10:58:20 crc kubenswrapper[4934]: I1002 10:58:20.862786 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1826925-c5fc-4e16-a51b-da8e36c6ffcd" containerName="extract-content" Oct 02 10:58:20 crc kubenswrapper[4934]: E1002 10:58:20.862802 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1826925-c5fc-4e16-a51b-da8e36c6ffcd" containerName="extract-utilities" Oct 02 10:58:20 crc kubenswrapper[4934]: I1002 10:58:20.862811 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1826925-c5fc-4e16-a51b-da8e36c6ffcd" containerName="extract-utilities" Oct 02 10:58:20 crc kubenswrapper[4934]: E1002 10:58:20.862837 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1826925-c5fc-4e16-a51b-da8e36c6ffcd" containerName="registry-server" Oct 02 10:58:20 crc kubenswrapper[4934]: I1002 10:58:20.862845 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1826925-c5fc-4e16-a51b-da8e36c6ffcd" containerName="registry-server" Oct 02 10:58:20 crc kubenswrapper[4934]: I1002 10:58:20.863028 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1826925-c5fc-4e16-a51b-da8e36c6ffcd" containerName="registry-server" Oct 02 10:58:20 crc kubenswrapper[4934]: I1002 10:58:20.864340 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7677s" Oct 02 10:58:20 crc kubenswrapper[4934]: I1002 10:58:20.874167 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7677s"] Oct 02 10:58:20 crc kubenswrapper[4934]: I1002 10:58:20.966472 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed-catalog-content\") pod \"certified-operators-7677s\" (UID: \"1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed\") " pod="openshift-marketplace/certified-operators-7677s" Oct 02 10:58:20 crc kubenswrapper[4934]: I1002 10:58:20.966554 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9dxj\" (UniqueName: \"kubernetes.io/projected/1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed-kube-api-access-d9dxj\") pod \"certified-operators-7677s\" (UID: \"1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed\") " pod="openshift-marketplace/certified-operators-7677s" Oct 02 10:58:20 crc kubenswrapper[4934]: I1002 10:58:20.966619 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed-utilities\") pod \"certified-operators-7677s\" (UID: \"1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed\") " pod="openshift-marketplace/certified-operators-7677s" Oct 02 10:58:21 crc kubenswrapper[4934]: I1002 10:58:21.067429 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed-catalog-content\") pod \"certified-operators-7677s\" (UID: \"1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed\") " pod="openshift-marketplace/certified-operators-7677s" Oct 02 10:58:21 crc kubenswrapper[4934]: I1002 10:58:21.067512 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9dxj\" (UniqueName: \"kubernetes.io/projected/1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed-kube-api-access-d9dxj\") pod \"certified-operators-7677s\" (UID: \"1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed\") " pod="openshift-marketplace/certified-operators-7677s" Oct 02 10:58:21 crc kubenswrapper[4934]: I1002 10:58:21.067566 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed-utilities\") pod \"certified-operators-7677s\" (UID: \"1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed\") " pod="openshift-marketplace/certified-operators-7677s" Oct 02 10:58:21 crc kubenswrapper[4934]: I1002 10:58:21.068213 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed-utilities\") pod \"certified-operators-7677s\" (UID: \"1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed\") " pod="openshift-marketplace/certified-operators-7677s" Oct 02 10:58:21 crc kubenswrapper[4934]: I1002 10:58:21.068339 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed-catalog-content\") pod \"certified-operators-7677s\" (UID: \"1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed\") " pod="openshift-marketplace/certified-operators-7677s" Oct 02 10:58:21 crc kubenswrapper[4934]: I1002 10:58:21.092598 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9dxj\" (UniqueName: \"kubernetes.io/projected/1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed-kube-api-access-d9dxj\") pod \"certified-operators-7677s\" (UID: \"1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed\") " pod="openshift-marketplace/certified-operators-7677s" Oct 02 10:58:21 crc kubenswrapper[4934]: I1002 10:58:21.190768 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7677s" Oct 02 10:58:21 crc kubenswrapper[4934]: I1002 10:58:21.634021 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7677s"] Oct 02 10:58:21 crc kubenswrapper[4934]: W1002 10:58:21.639079 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b2da8e7_e09d_48cc_a8d2_ab433e32a0ed.slice/crio-4c4b2f6ce0ea8781a4d6ea2474f3e84d30d11b63754df4943d1586fbe43a0965 WatchSource:0}: Error finding container 4c4b2f6ce0ea8781a4d6ea2474f3e84d30d11b63754df4943d1586fbe43a0965: Status 404 returned error can't find the container with id 4c4b2f6ce0ea8781a4d6ea2474f3e84d30d11b63754df4943d1586fbe43a0965 Oct 02 10:58:21 crc kubenswrapper[4934]: I1002 10:58:21.971115 4934 generic.go:334] "Generic (PLEG): container finished" podID="1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed" containerID="174a756643ab5c678c57ee528c59e4d4e6c94916efedb3ab33e871957cd47f30" exitCode=0 Oct 02 10:58:21 crc kubenswrapper[4934]: I1002 10:58:21.971163 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7677s" event={"ID":"1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed","Type":"ContainerDied","Data":"174a756643ab5c678c57ee528c59e4d4e6c94916efedb3ab33e871957cd47f30"} Oct 02 10:58:21 crc kubenswrapper[4934]: I1002 10:58:21.971197 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7677s" event={"ID":"1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed","Type":"ContainerStarted","Data":"4c4b2f6ce0ea8781a4d6ea2474f3e84d30d11b63754df4943d1586fbe43a0965"} Oct 02 10:58:21 crc kubenswrapper[4934]: I1002 10:58:21.973338 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 10:58:22 crc kubenswrapper[4934]: I1002 10:58:22.981378 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7677s" event={"ID":"1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed","Type":"ContainerStarted","Data":"60b03a1075013ee3240e21aad131bf70851a65b249729c5a09e43a61555db9b4"} Oct 02 10:58:23 crc kubenswrapper[4934]: I1002 10:58:23.989828 4934 generic.go:334] "Generic (PLEG): container finished" podID="1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed" containerID="60b03a1075013ee3240e21aad131bf70851a65b249729c5a09e43a61555db9b4" exitCode=0 Oct 02 10:58:23 crc kubenswrapper[4934]: I1002 10:58:23.989899 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7677s" event={"ID":"1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed","Type":"ContainerDied","Data":"60b03a1075013ee3240e21aad131bf70851a65b249729c5a09e43a61555db9b4"} Oct 02 10:58:25 crc kubenswrapper[4934]: I1002 10:58:24.999738 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7677s" event={"ID":"1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed","Type":"ContainerStarted","Data":"842ffece21737da86559791c45b3032f858b535e233c5d44bf1d9d4990f50ed9"} Oct 02 10:58:25 crc kubenswrapper[4934]: I1002 10:58:25.022650 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7677s" podStartSLOduration=2.5567589379999998 podStartE2EDuration="5.022633461s" podCreationTimestamp="2025-10-02 10:58:20 +0000 UTC" firstStartedPulling="2025-10-02 10:58:21.972975657 +0000 UTC m=+4173.725617189" lastFinishedPulling="2025-10-02 10:58:24.43885019 +0000 UTC m=+4176.191491712" observedRunningTime="2025-10-02 10:58:25.021412147 +0000 UTC m=+4176.774053669" watchObservedRunningTime="2025-10-02 10:58:25.022633461 +0000 UTC m=+4176.775274983" Oct 02 10:58:31 crc kubenswrapper[4934]: I1002 10:58:31.191663 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7677s" Oct 02 10:58:31 crc kubenswrapper[4934]: I1002 10:58:31.192102 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7677s" Oct 02 10:58:31 crc kubenswrapper[4934]: I1002 10:58:31.243231 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7677s" Oct 02 10:58:32 crc kubenswrapper[4934]: I1002 10:58:32.116620 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7677s" Oct 02 10:58:32 crc kubenswrapper[4934]: I1002 10:58:32.173822 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7677s"] Oct 02 10:58:34 crc kubenswrapper[4934]: I1002 10:58:34.075021 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7677s" podUID="1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed" containerName="registry-server" containerID="cri-o://842ffece21737da86559791c45b3032f858b535e233c5d44bf1d9d4990f50ed9" gracePeriod=2 Oct 02 10:58:34 crc kubenswrapper[4934]: I1002 10:58:34.464533 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7677s" Oct 02 10:58:34 crc kubenswrapper[4934]: I1002 10:58:34.579641 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed-catalog-content\") pod \"1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed\" (UID: \"1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed\") " Oct 02 10:58:34 crc kubenswrapper[4934]: I1002 10:58:34.579762 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9dxj\" (UniqueName: \"kubernetes.io/projected/1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed-kube-api-access-d9dxj\") pod \"1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed\" (UID: \"1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed\") " Oct 02 10:58:34 crc kubenswrapper[4934]: I1002 10:58:34.579888 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed-utilities\") pod \"1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed\" (UID: \"1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed\") " Oct 02 10:58:34 crc kubenswrapper[4934]: I1002 10:58:34.581148 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed-utilities" (OuterVolumeSpecName: "utilities") pod "1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed" (UID: "1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:58:34 crc kubenswrapper[4934]: I1002 10:58:34.588298 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed-kube-api-access-d9dxj" (OuterVolumeSpecName: "kube-api-access-d9dxj") pod "1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed" (UID: "1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed"). InnerVolumeSpecName "kube-api-access-d9dxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:58:34 crc kubenswrapper[4934]: I1002 10:58:34.634991 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed" (UID: "1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:58:34 crc kubenswrapper[4934]: I1002 10:58:34.681738 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:58:34 crc kubenswrapper[4934]: I1002 10:58:34.681777 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:58:34 crc kubenswrapper[4934]: I1002 10:58:34.681791 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9dxj\" (UniqueName: \"kubernetes.io/projected/1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed-kube-api-access-d9dxj\") on node \"crc\" DevicePath \"\"" Oct 02 10:58:35 crc kubenswrapper[4934]: I1002 10:58:35.086384 4934 generic.go:334] "Generic (PLEG): container finished" podID="1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed" containerID="842ffece21737da86559791c45b3032f858b535e233c5d44bf1d9d4990f50ed9" exitCode=0 Oct 02 10:58:35 crc kubenswrapper[4934]: I1002 10:58:35.086454 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7677s" Oct 02 10:58:35 crc kubenswrapper[4934]: I1002 10:58:35.086454 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7677s" event={"ID":"1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed","Type":"ContainerDied","Data":"842ffece21737da86559791c45b3032f858b535e233c5d44bf1d9d4990f50ed9"} Oct 02 10:58:35 crc kubenswrapper[4934]: I1002 10:58:35.086539 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7677s" event={"ID":"1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed","Type":"ContainerDied","Data":"4c4b2f6ce0ea8781a4d6ea2474f3e84d30d11b63754df4943d1586fbe43a0965"} Oct 02 10:58:35 crc kubenswrapper[4934]: I1002 10:58:35.086612 4934 scope.go:117] "RemoveContainer" containerID="842ffece21737da86559791c45b3032f858b535e233c5d44bf1d9d4990f50ed9" Oct 02 10:58:35 crc kubenswrapper[4934]: I1002 10:58:35.116635 4934 scope.go:117] "RemoveContainer" containerID="60b03a1075013ee3240e21aad131bf70851a65b249729c5a09e43a61555db9b4" Oct 02 10:58:35 crc kubenswrapper[4934]: I1002 10:58:35.125859 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7677s"] Oct 02 10:58:35 crc kubenswrapper[4934]: I1002 10:58:35.133261 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7677s"] Oct 02 10:58:35 crc kubenswrapper[4934]: I1002 10:58:35.155095 4934 scope.go:117] "RemoveContainer" containerID="174a756643ab5c678c57ee528c59e4d4e6c94916efedb3ab33e871957cd47f30" Oct 02 10:58:35 crc kubenswrapper[4934]: I1002 10:58:35.176194 4934 scope.go:117] "RemoveContainer" containerID="842ffece21737da86559791c45b3032f858b535e233c5d44bf1d9d4990f50ed9" Oct 02 10:58:35 crc kubenswrapper[4934]: E1002 10:58:35.176856 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"842ffece21737da86559791c45b3032f858b535e233c5d44bf1d9d4990f50ed9\": container with ID starting with 842ffece21737da86559791c45b3032f858b535e233c5d44bf1d9d4990f50ed9 not found: ID does not exist" containerID="842ffece21737da86559791c45b3032f858b535e233c5d44bf1d9d4990f50ed9" Oct 02 10:58:35 crc kubenswrapper[4934]: I1002 10:58:35.176887 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"842ffece21737da86559791c45b3032f858b535e233c5d44bf1d9d4990f50ed9"} err="failed to get container status \"842ffece21737da86559791c45b3032f858b535e233c5d44bf1d9d4990f50ed9\": rpc error: code = NotFound desc = could not find container \"842ffece21737da86559791c45b3032f858b535e233c5d44bf1d9d4990f50ed9\": container with ID starting with 842ffece21737da86559791c45b3032f858b535e233c5d44bf1d9d4990f50ed9 not found: ID does not exist" Oct 02 10:58:35 crc kubenswrapper[4934]: I1002 10:58:35.176910 4934 scope.go:117] "RemoveContainer" containerID="60b03a1075013ee3240e21aad131bf70851a65b249729c5a09e43a61555db9b4" Oct 02 10:58:35 crc kubenswrapper[4934]: E1002 10:58:35.177219 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60b03a1075013ee3240e21aad131bf70851a65b249729c5a09e43a61555db9b4\": container with ID starting with 60b03a1075013ee3240e21aad131bf70851a65b249729c5a09e43a61555db9b4 not found: ID does not exist" containerID="60b03a1075013ee3240e21aad131bf70851a65b249729c5a09e43a61555db9b4" Oct 02 10:58:35 crc kubenswrapper[4934]: I1002 10:58:35.177320 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60b03a1075013ee3240e21aad131bf70851a65b249729c5a09e43a61555db9b4"} err="failed to get container status \"60b03a1075013ee3240e21aad131bf70851a65b249729c5a09e43a61555db9b4\": rpc error: code = NotFound desc = could not find container \"60b03a1075013ee3240e21aad131bf70851a65b249729c5a09e43a61555db9b4\": container with ID starting with 60b03a1075013ee3240e21aad131bf70851a65b249729c5a09e43a61555db9b4 not found: ID does not exist" Oct 02 10:58:35 crc kubenswrapper[4934]: I1002 10:58:35.177418 4934 scope.go:117] "RemoveContainer" containerID="174a756643ab5c678c57ee528c59e4d4e6c94916efedb3ab33e871957cd47f30" Oct 02 10:58:35 crc kubenswrapper[4934]: E1002 10:58:35.178142 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"174a756643ab5c678c57ee528c59e4d4e6c94916efedb3ab33e871957cd47f30\": container with ID starting with 174a756643ab5c678c57ee528c59e4d4e6c94916efedb3ab33e871957cd47f30 not found: ID does not exist" containerID="174a756643ab5c678c57ee528c59e4d4e6c94916efedb3ab33e871957cd47f30" Oct 02 10:58:35 crc kubenswrapper[4934]: I1002 10:58:35.178188 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"174a756643ab5c678c57ee528c59e4d4e6c94916efedb3ab33e871957cd47f30"} err="failed to get container status \"174a756643ab5c678c57ee528c59e4d4e6c94916efedb3ab33e871957cd47f30\": rpc error: code = NotFound desc = could not find container \"174a756643ab5c678c57ee528c59e4d4e6c94916efedb3ab33e871957cd47f30\": container with ID starting with 174a756643ab5c678c57ee528c59e4d4e6c94916efedb3ab33e871957cd47f30 not found: ID does not exist" Oct 02 10:58:36 crc kubenswrapper[4934]: I1002 10:58:36.924463 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed" path="/var/lib/kubelet/pods/1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed/volumes" Oct 02 10:58:44 crc kubenswrapper[4934]: I1002 10:58:44.774087 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5tb57"] Oct 02 10:58:44 crc kubenswrapper[4934]: E1002 10:58:44.775060 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed" containerName="extract-utilities" Oct 02 10:58:44 crc kubenswrapper[4934]: I1002 10:58:44.775083 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed" containerName="extract-utilities" Oct 02 10:58:44 crc kubenswrapper[4934]: E1002 10:58:44.775101 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed" containerName="extract-content" Oct 02 10:58:44 crc kubenswrapper[4934]: I1002 10:58:44.775112 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed" containerName="extract-content" Oct 02 10:58:44 crc kubenswrapper[4934]: E1002 10:58:44.775130 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed" containerName="registry-server" Oct 02 10:58:44 crc kubenswrapper[4934]: I1002 10:58:44.775139 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed" containerName="registry-server" Oct 02 10:58:44 crc kubenswrapper[4934]: I1002 10:58:44.775344 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b2da8e7-e09d-48cc-a8d2-ab433e32a0ed" containerName="registry-server" Oct 02 10:58:44 crc kubenswrapper[4934]: I1002 10:58:44.777284 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5tb57" Oct 02 10:58:44 crc kubenswrapper[4934]: I1002 10:58:44.794554 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5tb57"] Oct 02 10:58:44 crc kubenswrapper[4934]: I1002 10:58:44.924209 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n879w\" (UniqueName: \"kubernetes.io/projected/c3846565-6bd1-412f-8425-a32e7d1e9001-kube-api-access-n879w\") pod \"redhat-operators-5tb57\" (UID: \"c3846565-6bd1-412f-8425-a32e7d1e9001\") " pod="openshift-marketplace/redhat-operators-5tb57" Oct 02 10:58:44 crc kubenswrapper[4934]: I1002 10:58:44.924301 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3846565-6bd1-412f-8425-a32e7d1e9001-utilities\") pod \"redhat-operators-5tb57\" (UID: \"c3846565-6bd1-412f-8425-a32e7d1e9001\") " pod="openshift-marketplace/redhat-operators-5tb57" Oct 02 10:58:44 crc kubenswrapper[4934]: I1002 10:58:44.924331 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3846565-6bd1-412f-8425-a32e7d1e9001-catalog-content\") pod \"redhat-operators-5tb57\" (UID: \"c3846565-6bd1-412f-8425-a32e7d1e9001\") " pod="openshift-marketplace/redhat-operators-5tb57" Oct 02 10:58:45 crc kubenswrapper[4934]: I1002 10:58:45.026308 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n879w\" (UniqueName: \"kubernetes.io/projected/c3846565-6bd1-412f-8425-a32e7d1e9001-kube-api-access-n879w\") pod \"redhat-operators-5tb57\" (UID: \"c3846565-6bd1-412f-8425-a32e7d1e9001\") " pod="openshift-marketplace/redhat-operators-5tb57" Oct 02 10:58:45 crc kubenswrapper[4934]: I1002 10:58:45.026444 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3846565-6bd1-412f-8425-a32e7d1e9001-utilities\") pod \"redhat-operators-5tb57\" (UID: \"c3846565-6bd1-412f-8425-a32e7d1e9001\") " pod="openshift-marketplace/redhat-operators-5tb57" Oct 02 10:58:45 crc kubenswrapper[4934]: I1002 10:58:45.026483 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3846565-6bd1-412f-8425-a32e7d1e9001-catalog-content\") pod \"redhat-operators-5tb57\" (UID: \"c3846565-6bd1-412f-8425-a32e7d1e9001\") " pod="openshift-marketplace/redhat-operators-5tb57" Oct 02 10:58:45 crc kubenswrapper[4934]: I1002 10:58:45.026996 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3846565-6bd1-412f-8425-a32e7d1e9001-utilities\") pod \"redhat-operators-5tb57\" (UID: \"c3846565-6bd1-412f-8425-a32e7d1e9001\") " pod="openshift-marketplace/redhat-operators-5tb57" Oct 02 10:58:45 crc kubenswrapper[4934]: I1002 10:58:45.027102 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3846565-6bd1-412f-8425-a32e7d1e9001-catalog-content\") pod \"redhat-operators-5tb57\" (UID: \"c3846565-6bd1-412f-8425-a32e7d1e9001\") " pod="openshift-marketplace/redhat-operators-5tb57" Oct 02 10:58:45 crc kubenswrapper[4934]: I1002 10:58:45.055665 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n879w\" (UniqueName: \"kubernetes.io/projected/c3846565-6bd1-412f-8425-a32e7d1e9001-kube-api-access-n879w\") pod \"redhat-operators-5tb57\" (UID: \"c3846565-6bd1-412f-8425-a32e7d1e9001\") " pod="openshift-marketplace/redhat-operators-5tb57" Oct 02 10:58:45 crc kubenswrapper[4934]: I1002 10:58:45.136635 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5tb57" Oct 02 10:58:45 crc kubenswrapper[4934]: I1002 10:58:45.570982 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5tb57"] Oct 02 10:58:46 crc kubenswrapper[4934]: I1002 10:58:46.196298 4934 generic.go:334] "Generic (PLEG): container finished" podID="c3846565-6bd1-412f-8425-a32e7d1e9001" containerID="fa464a3f4fc3bd47c1ae7e6b40b41f097f5d37ac33ab4bb18a2c2c1f13ca04e0" exitCode=0 Oct 02 10:58:46 crc kubenswrapper[4934]: I1002 10:58:46.196541 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5tb57" event={"ID":"c3846565-6bd1-412f-8425-a32e7d1e9001","Type":"ContainerDied","Data":"fa464a3f4fc3bd47c1ae7e6b40b41f097f5d37ac33ab4bb18a2c2c1f13ca04e0"} Oct 02 10:58:46 crc kubenswrapper[4934]: I1002 10:58:46.196852 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5tb57" event={"ID":"c3846565-6bd1-412f-8425-a32e7d1e9001","Type":"ContainerStarted","Data":"413e4400fd530f95a642dff003ecfb0b5aaed54bd62ae2746bdf97fb3da66e08"} Oct 02 10:58:47 crc kubenswrapper[4934]: I1002 10:58:47.207067 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5tb57" event={"ID":"c3846565-6bd1-412f-8425-a32e7d1e9001","Type":"ContainerStarted","Data":"f15f9ff3f145fbdcefb0166269d1181f3127e7de5b0c649df2e57c49d7a6e1fb"} Oct 02 10:58:48 crc kubenswrapper[4934]: I1002 10:58:48.216797 4934 generic.go:334] "Generic (PLEG): container finished" podID="c3846565-6bd1-412f-8425-a32e7d1e9001" containerID="f15f9ff3f145fbdcefb0166269d1181f3127e7de5b0c649df2e57c49d7a6e1fb" exitCode=0 Oct 02 10:58:48 crc kubenswrapper[4934]: I1002 10:58:48.216950 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5tb57" event={"ID":"c3846565-6bd1-412f-8425-a32e7d1e9001","Type":"ContainerDied","Data":"f15f9ff3f145fbdcefb0166269d1181f3127e7de5b0c649df2e57c49d7a6e1fb"} Oct 02 10:58:49 crc kubenswrapper[4934]: I1002 10:58:49.226944 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5tb57" event={"ID":"c3846565-6bd1-412f-8425-a32e7d1e9001","Type":"ContainerStarted","Data":"d60a583f402a2fca3499e0792ca9af0e2b01bd731c6e1099bb555c565d995c8c"} Oct 02 10:58:49 crc kubenswrapper[4934]: I1002 10:58:49.252546 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5tb57" podStartSLOduration=2.80528847 podStartE2EDuration="5.252527772s" podCreationTimestamp="2025-10-02 10:58:44 +0000 UTC" firstStartedPulling="2025-10-02 10:58:46.198840657 +0000 UTC m=+4197.951482249" lastFinishedPulling="2025-10-02 10:58:48.646079999 +0000 UTC m=+4200.398721551" observedRunningTime="2025-10-02 10:58:49.248552473 +0000 UTC m=+4201.001194025" watchObservedRunningTime="2025-10-02 10:58:49.252527772 +0000 UTC m=+4201.005169304" Oct 02 10:58:55 crc kubenswrapper[4934]: I1002 10:58:55.137088 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5tb57" Oct 02 10:58:55 crc kubenswrapper[4934]: I1002 10:58:55.137664 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5tb57" Oct 02 10:58:55 crc kubenswrapper[4934]: I1002 10:58:55.185947 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5tb57" Oct 02 10:58:55 crc kubenswrapper[4934]: I1002 10:58:55.316061 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5tb57" Oct 02 10:58:55 crc kubenswrapper[4934]: I1002 10:58:55.414902 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5tb57"] Oct 02 10:58:57 crc kubenswrapper[4934]: I1002 10:58:57.290161 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5tb57" podUID="c3846565-6bd1-412f-8425-a32e7d1e9001" containerName="registry-server" containerID="cri-o://d60a583f402a2fca3499e0792ca9af0e2b01bd731c6e1099bb555c565d995c8c" gracePeriod=2 Oct 02 10:58:57 crc kubenswrapper[4934]: I1002 10:58:57.670247 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5tb57" Oct 02 10:58:57 crc kubenswrapper[4934]: I1002 10:58:57.724788 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3846565-6bd1-412f-8425-a32e7d1e9001-catalog-content\") pod \"c3846565-6bd1-412f-8425-a32e7d1e9001\" (UID: \"c3846565-6bd1-412f-8425-a32e7d1e9001\") " Oct 02 10:58:57 crc kubenswrapper[4934]: I1002 10:58:57.724866 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n879w\" (UniqueName: \"kubernetes.io/projected/c3846565-6bd1-412f-8425-a32e7d1e9001-kube-api-access-n879w\") pod \"c3846565-6bd1-412f-8425-a32e7d1e9001\" (UID: \"c3846565-6bd1-412f-8425-a32e7d1e9001\") " Oct 02 10:58:57 crc kubenswrapper[4934]: I1002 10:58:57.724980 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3846565-6bd1-412f-8425-a32e7d1e9001-utilities\") pod \"c3846565-6bd1-412f-8425-a32e7d1e9001\" (UID: \"c3846565-6bd1-412f-8425-a32e7d1e9001\") " Oct 02 10:58:57 crc kubenswrapper[4934]: I1002 10:58:57.726081 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3846565-6bd1-412f-8425-a32e7d1e9001-utilities" (OuterVolumeSpecName: "utilities") pod "c3846565-6bd1-412f-8425-a32e7d1e9001" (UID: "c3846565-6bd1-412f-8425-a32e7d1e9001"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:58:57 crc kubenswrapper[4934]: I1002 10:58:57.734887 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3846565-6bd1-412f-8425-a32e7d1e9001-kube-api-access-n879w" (OuterVolumeSpecName: "kube-api-access-n879w") pod "c3846565-6bd1-412f-8425-a32e7d1e9001" (UID: "c3846565-6bd1-412f-8425-a32e7d1e9001"). InnerVolumeSpecName "kube-api-access-n879w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:58:57 crc kubenswrapper[4934]: I1002 10:58:57.826710 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3846565-6bd1-412f-8425-a32e7d1e9001-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:58:57 crc kubenswrapper[4934]: I1002 10:58:57.826751 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n879w\" (UniqueName: \"kubernetes.io/projected/c3846565-6bd1-412f-8425-a32e7d1e9001-kube-api-access-n879w\") on node \"crc\" DevicePath \"\"" Oct 02 10:58:58 crc kubenswrapper[4934]: I1002 10:58:58.303610 4934 generic.go:334] "Generic (PLEG): container finished" podID="c3846565-6bd1-412f-8425-a32e7d1e9001" containerID="d60a583f402a2fca3499e0792ca9af0e2b01bd731c6e1099bb555c565d995c8c" exitCode=0 Oct 02 10:58:58 crc kubenswrapper[4934]: I1002 10:58:58.303691 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5tb57" Oct 02 10:58:58 crc kubenswrapper[4934]: I1002 10:58:58.303720 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5tb57" event={"ID":"c3846565-6bd1-412f-8425-a32e7d1e9001","Type":"ContainerDied","Data":"d60a583f402a2fca3499e0792ca9af0e2b01bd731c6e1099bb555c565d995c8c"} Oct 02 10:58:58 crc kubenswrapper[4934]: I1002 10:58:58.304257 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5tb57" event={"ID":"c3846565-6bd1-412f-8425-a32e7d1e9001","Type":"ContainerDied","Data":"413e4400fd530f95a642dff003ecfb0b5aaed54bd62ae2746bdf97fb3da66e08"} Oct 02 10:58:58 crc kubenswrapper[4934]: I1002 10:58:58.304302 4934 scope.go:117] "RemoveContainer" containerID="d60a583f402a2fca3499e0792ca9af0e2b01bd731c6e1099bb555c565d995c8c" Oct 02 10:58:58 crc kubenswrapper[4934]: I1002 10:58:58.333746 4934 scope.go:117] "RemoveContainer" containerID="f15f9ff3f145fbdcefb0166269d1181f3127e7de5b0c649df2e57c49d7a6e1fb" Oct 02 10:58:58 crc kubenswrapper[4934]: I1002 10:58:58.365527 4934 scope.go:117] "RemoveContainer" containerID="fa464a3f4fc3bd47c1ae7e6b40b41f097f5d37ac33ab4bb18a2c2c1f13ca04e0" Oct 02 10:58:58 crc kubenswrapper[4934]: I1002 10:58:58.404202 4934 scope.go:117] "RemoveContainer" containerID="d60a583f402a2fca3499e0792ca9af0e2b01bd731c6e1099bb555c565d995c8c" Oct 02 10:58:58 crc kubenswrapper[4934]: E1002 10:58:58.405383 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d60a583f402a2fca3499e0792ca9af0e2b01bd731c6e1099bb555c565d995c8c\": container with ID starting with d60a583f402a2fca3499e0792ca9af0e2b01bd731c6e1099bb555c565d995c8c not found: ID does not exist" containerID="d60a583f402a2fca3499e0792ca9af0e2b01bd731c6e1099bb555c565d995c8c" Oct 02 10:58:58 crc kubenswrapper[4934]: I1002 10:58:58.405424 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d60a583f402a2fca3499e0792ca9af0e2b01bd731c6e1099bb555c565d995c8c"} err="failed to get container status \"d60a583f402a2fca3499e0792ca9af0e2b01bd731c6e1099bb555c565d995c8c\": rpc error: code = NotFound desc = could not find container \"d60a583f402a2fca3499e0792ca9af0e2b01bd731c6e1099bb555c565d995c8c\": container with ID starting with d60a583f402a2fca3499e0792ca9af0e2b01bd731c6e1099bb555c565d995c8c not found: ID does not exist" Oct 02 10:58:58 crc kubenswrapper[4934]: I1002 10:58:58.405450 4934 scope.go:117] "RemoveContainer" containerID="f15f9ff3f145fbdcefb0166269d1181f3127e7de5b0c649df2e57c49d7a6e1fb" Oct 02 10:58:58 crc kubenswrapper[4934]: E1002 10:58:58.405924 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f15f9ff3f145fbdcefb0166269d1181f3127e7de5b0c649df2e57c49d7a6e1fb\": container with ID starting with f15f9ff3f145fbdcefb0166269d1181f3127e7de5b0c649df2e57c49d7a6e1fb not found: ID does not exist" containerID="f15f9ff3f145fbdcefb0166269d1181f3127e7de5b0c649df2e57c49d7a6e1fb" Oct 02 10:58:58 crc kubenswrapper[4934]: I1002 10:58:58.405955 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f15f9ff3f145fbdcefb0166269d1181f3127e7de5b0c649df2e57c49d7a6e1fb"} err="failed to get container status \"f15f9ff3f145fbdcefb0166269d1181f3127e7de5b0c649df2e57c49d7a6e1fb\": rpc error: code = NotFound desc = could not find container \"f15f9ff3f145fbdcefb0166269d1181f3127e7de5b0c649df2e57c49d7a6e1fb\": container with ID starting with f15f9ff3f145fbdcefb0166269d1181f3127e7de5b0c649df2e57c49d7a6e1fb not found: ID does not exist" Oct 02 10:58:58 crc kubenswrapper[4934]: I1002 10:58:58.405974 4934 scope.go:117] "RemoveContainer" containerID="fa464a3f4fc3bd47c1ae7e6b40b41f097f5d37ac33ab4bb18a2c2c1f13ca04e0" Oct 02 10:58:58 crc kubenswrapper[4934]: E1002 10:58:58.406500 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa464a3f4fc3bd47c1ae7e6b40b41f097f5d37ac33ab4bb18a2c2c1f13ca04e0\": container with ID starting with fa464a3f4fc3bd47c1ae7e6b40b41f097f5d37ac33ab4bb18a2c2c1f13ca04e0 not found: ID does not exist" containerID="fa464a3f4fc3bd47c1ae7e6b40b41f097f5d37ac33ab4bb18a2c2c1f13ca04e0" Oct 02 10:58:58 crc kubenswrapper[4934]: I1002 10:58:58.406529 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa464a3f4fc3bd47c1ae7e6b40b41f097f5d37ac33ab4bb18a2c2c1f13ca04e0"} err="failed to get container status \"fa464a3f4fc3bd47c1ae7e6b40b41f097f5d37ac33ab4bb18a2c2c1f13ca04e0\": rpc error: code = NotFound desc = could not find container \"fa464a3f4fc3bd47c1ae7e6b40b41f097f5d37ac33ab4bb18a2c2c1f13ca04e0\": container with ID starting with fa464a3f4fc3bd47c1ae7e6b40b41f097f5d37ac33ab4bb18a2c2c1f13ca04e0 not found: ID does not exist" Oct 02 10:58:58 crc kubenswrapper[4934]: I1002 10:58:58.894719 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3846565-6bd1-412f-8425-a32e7d1e9001-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c3846565-6bd1-412f-8425-a32e7d1e9001" (UID: "c3846565-6bd1-412f-8425-a32e7d1e9001"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:58:58 crc kubenswrapper[4934]: I1002 10:58:58.946141 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3846565-6bd1-412f-8425-a32e7d1e9001-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:58:58 crc kubenswrapper[4934]: I1002 10:58:58.961921 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5tb57"] Oct 02 10:58:58 crc kubenswrapper[4934]: I1002 10:58:58.968228 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5tb57"] Oct 02 10:59:00 crc kubenswrapper[4934]: I1002 10:59:00.931165 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3846565-6bd1-412f-8425-a32e7d1e9001" path="/var/lib/kubelet/pods/c3846565-6bd1-412f-8425-a32e7d1e9001/volumes" Oct 02 10:59:08 crc kubenswrapper[4934]: I1002 10:59:08.440401 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:59:08 crc kubenswrapper[4934]: I1002 10:59:08.441392 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:59:14 crc kubenswrapper[4934]: I1002 10:59:14.070846 4934 scope.go:117] "RemoveContainer" containerID="7dcb97ee85f738e224a2b80adcf8cc9238632bdc5b8e2ebb242427461681b486" Oct 02 10:59:14 crc kubenswrapper[4934]: I1002 10:59:14.195225 4934 scope.go:117] "RemoveContainer" containerID="dcdbb7a09d7323ed1d9f664606b00099ab84379ee4e768568d43530fbd607ab9" Oct 02 10:59:38 crc kubenswrapper[4934]: I1002 10:59:38.439716 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:59:38 crc kubenswrapper[4934]: I1002 10:59:38.440519 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:00:00 crc kubenswrapper[4934]: I1002 11:00:00.152165 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323380-bvmrq"] Oct 02 11:00:00 crc kubenswrapper[4934]: E1002 11:00:00.153164 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3846565-6bd1-412f-8425-a32e7d1e9001" containerName="extract-utilities" Oct 02 11:00:00 crc kubenswrapper[4934]: I1002 11:00:00.153182 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3846565-6bd1-412f-8425-a32e7d1e9001" containerName="extract-utilities" Oct 02 11:00:00 crc kubenswrapper[4934]: E1002 11:00:00.153199 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3846565-6bd1-412f-8425-a32e7d1e9001" containerName="registry-server" Oct 02 11:00:00 crc kubenswrapper[4934]: I1002 11:00:00.153207 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3846565-6bd1-412f-8425-a32e7d1e9001" containerName="registry-server" Oct 02 11:00:00 crc kubenswrapper[4934]: E1002 11:00:00.153225 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3846565-6bd1-412f-8425-a32e7d1e9001" containerName="extract-content" Oct 02 11:00:00 crc kubenswrapper[4934]: I1002 11:00:00.153233 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3846565-6bd1-412f-8425-a32e7d1e9001" containerName="extract-content" Oct 02 11:00:00 crc kubenswrapper[4934]: I1002 11:00:00.153435 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3846565-6bd1-412f-8425-a32e7d1e9001" containerName="registry-server" Oct 02 11:00:00 crc kubenswrapper[4934]: I1002 11:00:00.154075 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-bvmrq" Oct 02 11:00:00 crc kubenswrapper[4934]: I1002 11:00:00.156406 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 11:00:00 crc kubenswrapper[4934]: I1002 11:00:00.156542 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 11:00:00 crc kubenswrapper[4934]: I1002 11:00:00.162206 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323380-bvmrq"] Oct 02 11:00:00 crc kubenswrapper[4934]: I1002 11:00:00.274436 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/567476d3-6832-4e0f-a0de-425a841fe1a9-secret-volume\") pod \"collect-profiles-29323380-bvmrq\" (UID: \"567476d3-6832-4e0f-a0de-425a841fe1a9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-bvmrq" Oct 02 11:00:00 crc kubenswrapper[4934]: I1002 11:00:00.274802 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/567476d3-6832-4e0f-a0de-425a841fe1a9-config-volume\") pod \"collect-profiles-29323380-bvmrq\" (UID: \"567476d3-6832-4e0f-a0de-425a841fe1a9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-bvmrq" Oct 02 11:00:00 crc kubenswrapper[4934]: I1002 11:00:00.274907 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84rcv\" (UniqueName: \"kubernetes.io/projected/567476d3-6832-4e0f-a0de-425a841fe1a9-kube-api-access-84rcv\") pod \"collect-profiles-29323380-bvmrq\" (UID: \"567476d3-6832-4e0f-a0de-425a841fe1a9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-bvmrq" Oct 02 11:00:00 crc kubenswrapper[4934]: I1002 11:00:00.376138 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/567476d3-6832-4e0f-a0de-425a841fe1a9-secret-volume\") pod \"collect-profiles-29323380-bvmrq\" (UID: \"567476d3-6832-4e0f-a0de-425a841fe1a9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-bvmrq" Oct 02 11:00:00 crc kubenswrapper[4934]: I1002 11:00:00.376506 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/567476d3-6832-4e0f-a0de-425a841fe1a9-config-volume\") pod \"collect-profiles-29323380-bvmrq\" (UID: \"567476d3-6832-4e0f-a0de-425a841fe1a9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-bvmrq" Oct 02 11:00:00 crc kubenswrapper[4934]: I1002 11:00:00.376652 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84rcv\" (UniqueName: \"kubernetes.io/projected/567476d3-6832-4e0f-a0de-425a841fe1a9-kube-api-access-84rcv\") pod \"collect-profiles-29323380-bvmrq\" (UID: \"567476d3-6832-4e0f-a0de-425a841fe1a9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-bvmrq" Oct 02 11:00:00 crc kubenswrapper[4934]: I1002 11:00:00.378687 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/567476d3-6832-4e0f-a0de-425a841fe1a9-config-volume\") pod \"collect-profiles-29323380-bvmrq\" (UID: \"567476d3-6832-4e0f-a0de-425a841fe1a9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-bvmrq" Oct 02 11:00:00 crc kubenswrapper[4934]: I1002 11:00:00.383325 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/567476d3-6832-4e0f-a0de-425a841fe1a9-secret-volume\") pod \"collect-profiles-29323380-bvmrq\" (UID: \"567476d3-6832-4e0f-a0de-425a841fe1a9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-bvmrq" Oct 02 11:00:00 crc kubenswrapper[4934]: I1002 11:00:00.394420 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84rcv\" (UniqueName: \"kubernetes.io/projected/567476d3-6832-4e0f-a0de-425a841fe1a9-kube-api-access-84rcv\") pod \"collect-profiles-29323380-bvmrq\" (UID: \"567476d3-6832-4e0f-a0de-425a841fe1a9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-bvmrq" Oct 02 11:00:00 crc kubenswrapper[4934]: I1002 11:00:00.478075 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-bvmrq" Oct 02 11:00:00 crc kubenswrapper[4934]: I1002 11:00:00.892669 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323380-bvmrq"] Oct 02 11:00:01 crc kubenswrapper[4934]: I1002 11:00:01.859988 4934 generic.go:334] "Generic (PLEG): container finished" podID="567476d3-6832-4e0f-a0de-425a841fe1a9" containerID="4dd414acb6446c3ca64effc203d3bd860629607e7a8b40f52bd03f63d643ff8e" exitCode=0 Oct 02 11:00:01 crc kubenswrapper[4934]: I1002 11:00:01.860103 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-bvmrq" event={"ID":"567476d3-6832-4e0f-a0de-425a841fe1a9","Type":"ContainerDied","Data":"4dd414acb6446c3ca64effc203d3bd860629607e7a8b40f52bd03f63d643ff8e"} Oct 02 11:00:01 crc kubenswrapper[4934]: I1002 11:00:01.860350 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-bvmrq" event={"ID":"567476d3-6832-4e0f-a0de-425a841fe1a9","Type":"ContainerStarted","Data":"70f749f60d63ed19d213311a4af84dc2a1bc67cb7c721a26563e779d0747c057"} Oct 02 11:00:03 crc kubenswrapper[4934]: I1002 11:00:03.142960 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-bvmrq" Oct 02 11:00:03 crc kubenswrapper[4934]: I1002 11:00:03.323003 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/567476d3-6832-4e0f-a0de-425a841fe1a9-config-volume\") pod \"567476d3-6832-4e0f-a0de-425a841fe1a9\" (UID: \"567476d3-6832-4e0f-a0de-425a841fe1a9\") " Oct 02 11:00:03 crc kubenswrapper[4934]: I1002 11:00:03.323151 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/567476d3-6832-4e0f-a0de-425a841fe1a9-secret-volume\") pod \"567476d3-6832-4e0f-a0de-425a841fe1a9\" (UID: \"567476d3-6832-4e0f-a0de-425a841fe1a9\") " Oct 02 11:00:03 crc kubenswrapper[4934]: I1002 11:00:03.323199 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84rcv\" (UniqueName: \"kubernetes.io/projected/567476d3-6832-4e0f-a0de-425a841fe1a9-kube-api-access-84rcv\") pod \"567476d3-6832-4e0f-a0de-425a841fe1a9\" (UID: \"567476d3-6832-4e0f-a0de-425a841fe1a9\") " Oct 02 11:00:03 crc kubenswrapper[4934]: I1002 11:00:03.324219 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/567476d3-6832-4e0f-a0de-425a841fe1a9-config-volume" (OuterVolumeSpecName: "config-volume") pod "567476d3-6832-4e0f-a0de-425a841fe1a9" (UID: "567476d3-6832-4e0f-a0de-425a841fe1a9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:00:03 crc kubenswrapper[4934]: I1002 11:00:03.329259 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/567476d3-6832-4e0f-a0de-425a841fe1a9-kube-api-access-84rcv" (OuterVolumeSpecName: "kube-api-access-84rcv") pod "567476d3-6832-4e0f-a0de-425a841fe1a9" (UID: "567476d3-6832-4e0f-a0de-425a841fe1a9"). InnerVolumeSpecName "kube-api-access-84rcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:00:03 crc kubenswrapper[4934]: I1002 11:00:03.329268 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/567476d3-6832-4e0f-a0de-425a841fe1a9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "567476d3-6832-4e0f-a0de-425a841fe1a9" (UID: "567476d3-6832-4e0f-a0de-425a841fe1a9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:00:03 crc kubenswrapper[4934]: I1002 11:00:03.424986 4934 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/567476d3-6832-4e0f-a0de-425a841fe1a9-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 11:00:03 crc kubenswrapper[4934]: I1002 11:00:03.425013 4934 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/567476d3-6832-4e0f-a0de-425a841fe1a9-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 11:00:03 crc kubenswrapper[4934]: I1002 11:00:03.425024 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84rcv\" (UniqueName: \"kubernetes.io/projected/567476d3-6832-4e0f-a0de-425a841fe1a9-kube-api-access-84rcv\") on node \"crc\" DevicePath \"\"" Oct 02 11:00:03 crc kubenswrapper[4934]: I1002 11:00:03.878667 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-bvmrq" event={"ID":"567476d3-6832-4e0f-a0de-425a841fe1a9","Type":"ContainerDied","Data":"70f749f60d63ed19d213311a4af84dc2a1bc67cb7c721a26563e779d0747c057"} Oct 02 11:00:03 crc kubenswrapper[4934]: I1002 11:00:03.878726 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70f749f60d63ed19d213311a4af84dc2a1bc67cb7c721a26563e779d0747c057" Oct 02 11:00:03 crc kubenswrapper[4934]: I1002 11:00:03.879193 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323380-bvmrq" Oct 02 11:00:04 crc kubenswrapper[4934]: I1002 11:00:04.225423 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323335-9gjf2"] Oct 02 11:00:04 crc kubenswrapper[4934]: I1002 11:00:04.230975 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323335-9gjf2"] Oct 02 11:00:04 crc kubenswrapper[4934]: I1002 11:00:04.923091 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4774f929-979f-4e63-86d2-f01df01c7334" path="/var/lib/kubelet/pods/4774f929-979f-4e63-86d2-f01df01c7334/volumes" Oct 02 11:00:08 crc kubenswrapper[4934]: I1002 11:00:08.439782 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:00:08 crc kubenswrapper[4934]: I1002 11:00:08.440647 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:00:08 crc kubenswrapper[4934]: I1002 11:00:08.440727 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 11:00:08 crc kubenswrapper[4934]: I1002 11:00:08.442076 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39"} pod="openshift-machine-config-operator/machine-config-daemon-djh5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 11:00:08 crc kubenswrapper[4934]: I1002 11:00:08.442197 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" containerID="cri-o://ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39" gracePeriod=600 Oct 02 11:00:08 crc kubenswrapper[4934]: E1002 11:00:08.798239 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:00:08 crc kubenswrapper[4934]: I1002 11:00:08.921098 4934 generic.go:334] "Generic (PLEG): container finished" podID="71db06ef-05b0-4f58-b251-b27117a8500a" containerID="ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39" exitCode=0 Oct 02 11:00:08 crc kubenswrapper[4934]: I1002 11:00:08.924265 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerDied","Data":"ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39"} Oct 02 11:00:08 crc kubenswrapper[4934]: I1002 11:00:08.924361 4934 scope.go:117] "RemoveContainer" containerID="acd6fc421b82d19f0da3591bd2215c6dd60b927b95d1597f6b8f779ce91c0f24" Oct 02 11:00:08 crc kubenswrapper[4934]: I1002 11:00:08.924723 4934 scope.go:117] "RemoveContainer" containerID="ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39" Oct 02 11:00:08 crc kubenswrapper[4934]: E1002 11:00:08.924978 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:00:14 crc kubenswrapper[4934]: I1002 11:00:14.258820 4934 scope.go:117] "RemoveContainer" containerID="a3b55e537f2b753ac52d41ae11574d8d156a736700427ae68c33484e9521f50b" Oct 02 11:00:14 crc kubenswrapper[4934]: I1002 11:00:14.293163 4934 scope.go:117] "RemoveContainer" containerID="d3a6fe0cb3b008bfe59c918adaf4b941558112628e74bbf977144b34d72190a3" Oct 02 11:00:19 crc kubenswrapper[4934]: I1002 11:00:19.912630 4934 scope.go:117] "RemoveContainer" containerID="ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39" Oct 02 11:00:19 crc kubenswrapper[4934]: E1002 11:00:19.913364 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:00:31 crc kubenswrapper[4934]: I1002 11:00:31.913364 4934 scope.go:117] "RemoveContainer" containerID="ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39" Oct 02 11:00:31 crc kubenswrapper[4934]: E1002 11:00:31.914153 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:00:45 crc kubenswrapper[4934]: I1002 11:00:45.913059 4934 scope.go:117] "RemoveContainer" containerID="ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39" Oct 02 11:00:45 crc kubenswrapper[4934]: E1002 11:00:45.913680 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:00:56 crc kubenswrapper[4934]: I1002 11:00:56.913469 4934 scope.go:117] "RemoveContainer" containerID="ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39" Oct 02 11:00:56 crc kubenswrapper[4934]: E1002 11:00:56.915255 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:01:09 crc kubenswrapper[4934]: I1002 11:01:09.913847 4934 scope.go:117] "RemoveContainer" containerID="ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39" Oct 02 11:01:09 crc kubenswrapper[4934]: E1002 11:01:09.914892 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:01:23 crc kubenswrapper[4934]: I1002 11:01:23.913074 4934 scope.go:117] "RemoveContainer" containerID="ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39" Oct 02 11:01:23 crc kubenswrapper[4934]: E1002 11:01:23.913772 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:01:35 crc kubenswrapper[4934]: I1002 11:01:35.937771 4934 scope.go:117] "RemoveContainer" containerID="ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39" Oct 02 11:01:35 crc kubenswrapper[4934]: E1002 11:01:35.938812 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:01:49 crc kubenswrapper[4934]: I1002 11:01:49.914040 4934 scope.go:117] "RemoveContainer" containerID="ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39" Oct 02 11:01:49 crc kubenswrapper[4934]: E1002 11:01:49.914988 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:02:00 crc kubenswrapper[4934]: I1002 11:02:00.913291 4934 scope.go:117] "RemoveContainer" containerID="ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39" Oct 02 11:02:00 crc kubenswrapper[4934]: E1002 11:02:00.913946 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:02:11 crc kubenswrapper[4934]: I1002 11:02:11.913336 4934 scope.go:117] "RemoveContainer" containerID="ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39" Oct 02 11:02:11 crc kubenswrapper[4934]: E1002 11:02:11.914061 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:02:22 crc kubenswrapper[4934]: I1002 11:02:22.111278 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x9kgb"] Oct 02 11:02:22 crc kubenswrapper[4934]: E1002 11:02:22.112282 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="567476d3-6832-4e0f-a0de-425a841fe1a9" containerName="collect-profiles" Oct 02 11:02:22 crc kubenswrapper[4934]: I1002 11:02:22.112299 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="567476d3-6832-4e0f-a0de-425a841fe1a9" containerName="collect-profiles" Oct 02 11:02:22 crc kubenswrapper[4934]: I1002 11:02:22.112485 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="567476d3-6832-4e0f-a0de-425a841fe1a9" containerName="collect-profiles" Oct 02 11:02:22 crc kubenswrapper[4934]: I1002 11:02:22.113732 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x9kgb" Oct 02 11:02:22 crc kubenswrapper[4934]: I1002 11:02:22.129522 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x9kgb"] Oct 02 11:02:22 crc kubenswrapper[4934]: I1002 11:02:22.164418 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66l59\" (UniqueName: \"kubernetes.io/projected/2263d5aa-449b-448d-b567-cf24d9c15b74-kube-api-access-66l59\") pod \"redhat-marketplace-x9kgb\" (UID: \"2263d5aa-449b-448d-b567-cf24d9c15b74\") " pod="openshift-marketplace/redhat-marketplace-x9kgb" Oct 02 11:02:22 crc kubenswrapper[4934]: I1002 11:02:22.164514 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2263d5aa-449b-448d-b567-cf24d9c15b74-catalog-content\") pod \"redhat-marketplace-x9kgb\" (UID: \"2263d5aa-449b-448d-b567-cf24d9c15b74\") " pod="openshift-marketplace/redhat-marketplace-x9kgb" Oct 02 11:02:22 crc kubenswrapper[4934]: I1002 11:02:22.164740 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2263d5aa-449b-448d-b567-cf24d9c15b74-utilities\") pod \"redhat-marketplace-x9kgb\" (UID: \"2263d5aa-449b-448d-b567-cf24d9c15b74\") " pod="openshift-marketplace/redhat-marketplace-x9kgb" Oct 02 11:02:22 crc kubenswrapper[4934]: I1002 11:02:22.265852 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66l59\" (UniqueName: \"kubernetes.io/projected/2263d5aa-449b-448d-b567-cf24d9c15b74-kube-api-access-66l59\") pod \"redhat-marketplace-x9kgb\" (UID: \"2263d5aa-449b-448d-b567-cf24d9c15b74\") " pod="openshift-marketplace/redhat-marketplace-x9kgb" Oct 02 11:02:22 crc kubenswrapper[4934]: I1002 11:02:22.265932 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2263d5aa-449b-448d-b567-cf24d9c15b74-catalog-content\") pod \"redhat-marketplace-x9kgb\" (UID: \"2263d5aa-449b-448d-b567-cf24d9c15b74\") " pod="openshift-marketplace/redhat-marketplace-x9kgb" Oct 02 11:02:22 crc kubenswrapper[4934]: I1002 11:02:22.265984 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2263d5aa-449b-448d-b567-cf24d9c15b74-utilities\") pod \"redhat-marketplace-x9kgb\" (UID: \"2263d5aa-449b-448d-b567-cf24d9c15b74\") " pod="openshift-marketplace/redhat-marketplace-x9kgb" Oct 02 11:02:22 crc kubenswrapper[4934]: I1002 11:02:22.266592 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2263d5aa-449b-448d-b567-cf24d9c15b74-utilities\") pod \"redhat-marketplace-x9kgb\" (UID: \"2263d5aa-449b-448d-b567-cf24d9c15b74\") " pod="openshift-marketplace/redhat-marketplace-x9kgb" Oct 02 11:02:22 crc kubenswrapper[4934]: I1002 11:02:22.267198 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2263d5aa-449b-448d-b567-cf24d9c15b74-catalog-content\") pod \"redhat-marketplace-x9kgb\" (UID: \"2263d5aa-449b-448d-b567-cf24d9c15b74\") " pod="openshift-marketplace/redhat-marketplace-x9kgb" Oct 02 11:02:22 crc kubenswrapper[4934]: I1002 11:02:22.292498 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66l59\" (UniqueName: \"kubernetes.io/projected/2263d5aa-449b-448d-b567-cf24d9c15b74-kube-api-access-66l59\") pod \"redhat-marketplace-x9kgb\" (UID: \"2263d5aa-449b-448d-b567-cf24d9c15b74\") " pod="openshift-marketplace/redhat-marketplace-x9kgb" Oct 02 11:02:22 crc kubenswrapper[4934]: I1002 11:02:22.436493 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x9kgb" Oct 02 11:02:22 crc kubenswrapper[4934]: I1002 11:02:22.844671 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x9kgb"] Oct 02 11:02:22 crc kubenswrapper[4934]: I1002 11:02:22.941216 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x9kgb" event={"ID":"2263d5aa-449b-448d-b567-cf24d9c15b74","Type":"ContainerStarted","Data":"eb8583fdac1e1bc0bb54cf8355055937ce68cd3d60333a00128606782c0ef884"} Oct 02 11:02:23 crc kubenswrapper[4934]: I1002 11:02:23.913933 4934 scope.go:117] "RemoveContainer" containerID="ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39" Oct 02 11:02:23 crc kubenswrapper[4934]: E1002 11:02:23.914311 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:02:23 crc kubenswrapper[4934]: I1002 11:02:23.955193 4934 generic.go:334] "Generic (PLEG): container finished" podID="2263d5aa-449b-448d-b567-cf24d9c15b74" containerID="54b0f509b5f474286820ddfca33968eff879b5fb37ec8d5223bccb65237cbd3c" exitCode=0 Oct 02 11:02:23 crc kubenswrapper[4934]: I1002 11:02:23.955281 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x9kgb" event={"ID":"2263d5aa-449b-448d-b567-cf24d9c15b74","Type":"ContainerDied","Data":"54b0f509b5f474286820ddfca33968eff879b5fb37ec8d5223bccb65237cbd3c"} Oct 02 11:02:24 crc kubenswrapper[4934]: I1002 11:02:24.965024 4934 generic.go:334] "Generic (PLEG): container finished" podID="2263d5aa-449b-448d-b567-cf24d9c15b74" containerID="e767ae7ce42346eb78f39b07f66ea937c1bb76050dd16984b05dfdf0b76c7662" exitCode=0 Oct 02 11:02:24 crc kubenswrapper[4934]: I1002 11:02:24.965065 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x9kgb" event={"ID":"2263d5aa-449b-448d-b567-cf24d9c15b74","Type":"ContainerDied","Data":"e767ae7ce42346eb78f39b07f66ea937c1bb76050dd16984b05dfdf0b76c7662"} Oct 02 11:02:25 crc kubenswrapper[4934]: I1002 11:02:25.974258 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x9kgb" event={"ID":"2263d5aa-449b-448d-b567-cf24d9c15b74","Type":"ContainerStarted","Data":"a33a4f059ae8dfe92f0fe762155a986d4a91ab4fca09e8059a64f3426cd6bfc6"} Oct 02 11:02:25 crc kubenswrapper[4934]: I1002 11:02:25.992217 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x9kgb" podStartSLOduration=2.5304996170000003 podStartE2EDuration="3.99219801s" podCreationTimestamp="2025-10-02 11:02:22 +0000 UTC" firstStartedPulling="2025-10-02 11:02:23.95883236 +0000 UTC m=+4415.711473892" lastFinishedPulling="2025-10-02 11:02:25.420530763 +0000 UTC m=+4417.173172285" observedRunningTime="2025-10-02 11:02:25.988509498 +0000 UTC m=+4417.741151050" watchObservedRunningTime="2025-10-02 11:02:25.99219801 +0000 UTC m=+4417.744839532" Oct 02 11:02:32 crc kubenswrapper[4934]: I1002 11:02:32.438219 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x9kgb" Oct 02 11:02:32 crc kubenswrapper[4934]: I1002 11:02:32.439044 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x9kgb" Oct 02 11:02:32 crc kubenswrapper[4934]: I1002 11:02:32.480833 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x9kgb" Oct 02 11:02:33 crc kubenswrapper[4934]: I1002 11:02:33.071865 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x9kgb" Oct 02 11:02:33 crc kubenswrapper[4934]: I1002 11:02:33.111924 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x9kgb"] Oct 02 11:02:35 crc kubenswrapper[4934]: I1002 11:02:35.044385 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-x9kgb" podUID="2263d5aa-449b-448d-b567-cf24d9c15b74" containerName="registry-server" containerID="cri-o://a33a4f059ae8dfe92f0fe762155a986d4a91ab4fca09e8059a64f3426cd6bfc6" gracePeriod=2 Oct 02 11:02:35 crc kubenswrapper[4934]: I1002 11:02:35.496377 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x9kgb" Oct 02 11:02:35 crc kubenswrapper[4934]: I1002 11:02:35.663645 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2263d5aa-449b-448d-b567-cf24d9c15b74-utilities\") pod \"2263d5aa-449b-448d-b567-cf24d9c15b74\" (UID: \"2263d5aa-449b-448d-b567-cf24d9c15b74\") " Oct 02 11:02:35 crc kubenswrapper[4934]: I1002 11:02:35.663716 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66l59\" (UniqueName: \"kubernetes.io/projected/2263d5aa-449b-448d-b567-cf24d9c15b74-kube-api-access-66l59\") pod \"2263d5aa-449b-448d-b567-cf24d9c15b74\" (UID: \"2263d5aa-449b-448d-b567-cf24d9c15b74\") " Oct 02 11:02:35 crc kubenswrapper[4934]: I1002 11:02:35.664802 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2263d5aa-449b-448d-b567-cf24d9c15b74-catalog-content\") pod \"2263d5aa-449b-448d-b567-cf24d9c15b74\" (UID: \"2263d5aa-449b-448d-b567-cf24d9c15b74\") " Oct 02 11:02:35 crc kubenswrapper[4934]: I1002 11:02:35.664826 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2263d5aa-449b-448d-b567-cf24d9c15b74-utilities" (OuterVolumeSpecName: "utilities") pod "2263d5aa-449b-448d-b567-cf24d9c15b74" (UID: "2263d5aa-449b-448d-b567-cf24d9c15b74"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:02:35 crc kubenswrapper[4934]: I1002 11:02:35.670050 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2263d5aa-449b-448d-b567-cf24d9c15b74-kube-api-access-66l59" (OuterVolumeSpecName: "kube-api-access-66l59") pod "2263d5aa-449b-448d-b567-cf24d9c15b74" (UID: "2263d5aa-449b-448d-b567-cf24d9c15b74"). InnerVolumeSpecName "kube-api-access-66l59". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:02:35 crc kubenswrapper[4934]: I1002 11:02:35.677887 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2263d5aa-449b-448d-b567-cf24d9c15b74-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2263d5aa-449b-448d-b567-cf24d9c15b74" (UID: "2263d5aa-449b-448d-b567-cf24d9c15b74"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:02:35 crc kubenswrapper[4934]: I1002 11:02:35.766445 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66l59\" (UniqueName: \"kubernetes.io/projected/2263d5aa-449b-448d-b567-cf24d9c15b74-kube-api-access-66l59\") on node \"crc\" DevicePath \"\"" Oct 02 11:02:35 crc kubenswrapper[4934]: I1002 11:02:35.766842 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2263d5aa-449b-448d-b567-cf24d9c15b74-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:02:35 crc kubenswrapper[4934]: I1002 11:02:35.767013 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2263d5aa-449b-448d-b567-cf24d9c15b74-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:02:36 crc kubenswrapper[4934]: I1002 11:02:36.053430 4934 generic.go:334] "Generic (PLEG): container finished" podID="2263d5aa-449b-448d-b567-cf24d9c15b74" containerID="a33a4f059ae8dfe92f0fe762155a986d4a91ab4fca09e8059a64f3426cd6bfc6" exitCode=0 Oct 02 11:02:36 crc kubenswrapper[4934]: I1002 11:02:36.053484 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x9kgb" event={"ID":"2263d5aa-449b-448d-b567-cf24d9c15b74","Type":"ContainerDied","Data":"a33a4f059ae8dfe92f0fe762155a986d4a91ab4fca09e8059a64f3426cd6bfc6"} Oct 02 11:02:36 crc kubenswrapper[4934]: I1002 11:02:36.053552 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x9kgb" Oct 02 11:02:36 crc kubenswrapper[4934]: I1002 11:02:36.053625 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x9kgb" event={"ID":"2263d5aa-449b-448d-b567-cf24d9c15b74","Type":"ContainerDied","Data":"eb8583fdac1e1bc0bb54cf8355055937ce68cd3d60333a00128606782c0ef884"} Oct 02 11:02:36 crc kubenswrapper[4934]: I1002 11:02:36.053667 4934 scope.go:117] "RemoveContainer" containerID="a33a4f059ae8dfe92f0fe762155a986d4a91ab4fca09e8059a64f3426cd6bfc6" Oct 02 11:02:36 crc kubenswrapper[4934]: I1002 11:02:36.080453 4934 scope.go:117] "RemoveContainer" containerID="e767ae7ce42346eb78f39b07f66ea937c1bb76050dd16984b05dfdf0b76c7662" Oct 02 11:02:36 crc kubenswrapper[4934]: I1002 11:02:36.097099 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x9kgb"] Oct 02 11:02:36 crc kubenswrapper[4934]: I1002 11:02:36.101265 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-x9kgb"] Oct 02 11:02:36 crc kubenswrapper[4934]: I1002 11:02:36.119108 4934 scope.go:117] "RemoveContainer" containerID="54b0f509b5f474286820ddfca33968eff879b5fb37ec8d5223bccb65237cbd3c" Oct 02 11:02:36 crc kubenswrapper[4934]: I1002 11:02:36.136541 4934 scope.go:117] "RemoveContainer" containerID="a33a4f059ae8dfe92f0fe762155a986d4a91ab4fca09e8059a64f3426cd6bfc6" Oct 02 11:02:36 crc kubenswrapper[4934]: E1002 11:02:36.136973 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a33a4f059ae8dfe92f0fe762155a986d4a91ab4fca09e8059a64f3426cd6bfc6\": container with ID starting with a33a4f059ae8dfe92f0fe762155a986d4a91ab4fca09e8059a64f3426cd6bfc6 not found: ID does not exist" containerID="a33a4f059ae8dfe92f0fe762155a986d4a91ab4fca09e8059a64f3426cd6bfc6" Oct 02 11:02:36 crc kubenswrapper[4934]: I1002 11:02:36.137007 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a33a4f059ae8dfe92f0fe762155a986d4a91ab4fca09e8059a64f3426cd6bfc6"} err="failed to get container status \"a33a4f059ae8dfe92f0fe762155a986d4a91ab4fca09e8059a64f3426cd6bfc6\": rpc error: code = NotFound desc = could not find container \"a33a4f059ae8dfe92f0fe762155a986d4a91ab4fca09e8059a64f3426cd6bfc6\": container with ID starting with a33a4f059ae8dfe92f0fe762155a986d4a91ab4fca09e8059a64f3426cd6bfc6 not found: ID does not exist" Oct 02 11:02:36 crc kubenswrapper[4934]: I1002 11:02:36.137032 4934 scope.go:117] "RemoveContainer" containerID="e767ae7ce42346eb78f39b07f66ea937c1bb76050dd16984b05dfdf0b76c7662" Oct 02 11:02:36 crc kubenswrapper[4934]: E1002 11:02:36.137443 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e767ae7ce42346eb78f39b07f66ea937c1bb76050dd16984b05dfdf0b76c7662\": container with ID starting with e767ae7ce42346eb78f39b07f66ea937c1bb76050dd16984b05dfdf0b76c7662 not found: ID does not exist" containerID="e767ae7ce42346eb78f39b07f66ea937c1bb76050dd16984b05dfdf0b76c7662" Oct 02 11:02:36 crc kubenswrapper[4934]: I1002 11:02:36.137479 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e767ae7ce42346eb78f39b07f66ea937c1bb76050dd16984b05dfdf0b76c7662"} err="failed to get container status \"e767ae7ce42346eb78f39b07f66ea937c1bb76050dd16984b05dfdf0b76c7662\": rpc error: code = NotFound desc = could not find container \"e767ae7ce42346eb78f39b07f66ea937c1bb76050dd16984b05dfdf0b76c7662\": container with ID starting with e767ae7ce42346eb78f39b07f66ea937c1bb76050dd16984b05dfdf0b76c7662 not found: ID does not exist" Oct 02 11:02:36 crc kubenswrapper[4934]: I1002 11:02:36.137505 4934 scope.go:117] "RemoveContainer" containerID="54b0f509b5f474286820ddfca33968eff879b5fb37ec8d5223bccb65237cbd3c" Oct 02 11:02:36 crc kubenswrapper[4934]: E1002 11:02:36.137853 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54b0f509b5f474286820ddfca33968eff879b5fb37ec8d5223bccb65237cbd3c\": container with ID starting with 54b0f509b5f474286820ddfca33968eff879b5fb37ec8d5223bccb65237cbd3c not found: ID does not exist" containerID="54b0f509b5f474286820ddfca33968eff879b5fb37ec8d5223bccb65237cbd3c" Oct 02 11:02:36 crc kubenswrapper[4934]: I1002 11:02:36.137897 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54b0f509b5f474286820ddfca33968eff879b5fb37ec8d5223bccb65237cbd3c"} err="failed to get container status \"54b0f509b5f474286820ddfca33968eff879b5fb37ec8d5223bccb65237cbd3c\": rpc error: code = NotFound desc = could not find container \"54b0f509b5f474286820ddfca33968eff879b5fb37ec8d5223bccb65237cbd3c\": container with ID starting with 54b0f509b5f474286820ddfca33968eff879b5fb37ec8d5223bccb65237cbd3c not found: ID does not exist" Oct 02 11:02:36 crc kubenswrapper[4934]: I1002 11:02:36.922085 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2263d5aa-449b-448d-b567-cf24d9c15b74" path="/var/lib/kubelet/pods/2263d5aa-449b-448d-b567-cf24d9c15b74/volumes" Oct 02 11:02:37 crc kubenswrapper[4934]: I1002 11:02:37.913978 4934 scope.go:117] "RemoveContainer" containerID="ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39" Oct 02 11:02:37 crc kubenswrapper[4934]: E1002 11:02:37.914877 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:02:49 crc kubenswrapper[4934]: I1002 11:02:49.912991 4934 scope.go:117] "RemoveContainer" containerID="ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39" Oct 02 11:02:49 crc kubenswrapper[4934]: E1002 11:02:49.914382 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:03:00 crc kubenswrapper[4934]: I1002 11:03:00.912776 4934 scope.go:117] "RemoveContainer" containerID="ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39" Oct 02 11:03:00 crc kubenswrapper[4934]: E1002 11:03:00.913463 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:03:13 crc kubenswrapper[4934]: I1002 11:03:13.912698 4934 scope.go:117] "RemoveContainer" containerID="ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39" Oct 02 11:03:13 crc kubenswrapper[4934]: E1002 11:03:13.914430 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:03:27 crc kubenswrapper[4934]: I1002 11:03:27.913003 4934 scope.go:117] "RemoveContainer" containerID="ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39" Oct 02 11:03:27 crc kubenswrapper[4934]: E1002 11:03:27.913870 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:03:41 crc kubenswrapper[4934]: I1002 11:03:41.914650 4934 scope.go:117] "RemoveContainer" containerID="ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39" Oct 02 11:03:41 crc kubenswrapper[4934]: E1002 11:03:41.915267 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:03:54 crc kubenswrapper[4934]: I1002 11:03:54.913261 4934 scope.go:117] "RemoveContainer" containerID="ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39" Oct 02 11:03:54 crc kubenswrapper[4934]: E1002 11:03:54.913947 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:04:08 crc kubenswrapper[4934]: I1002 11:04:08.917231 4934 scope.go:117] "RemoveContainer" containerID="ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39" Oct 02 11:04:08 crc kubenswrapper[4934]: E1002 11:04:08.917993 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:04:13 crc kubenswrapper[4934]: I1002 11:04:13.889378 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hcscn"] Oct 02 11:04:13 crc kubenswrapper[4934]: E1002 11:04:13.890098 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2263d5aa-449b-448d-b567-cf24d9c15b74" containerName="registry-server" Oct 02 11:04:13 crc kubenswrapper[4934]: I1002 11:04:13.890110 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="2263d5aa-449b-448d-b567-cf24d9c15b74" containerName="registry-server" Oct 02 11:04:13 crc kubenswrapper[4934]: E1002 11:04:13.890138 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2263d5aa-449b-448d-b567-cf24d9c15b74" containerName="extract-content" Oct 02 11:04:13 crc kubenswrapper[4934]: I1002 11:04:13.890144 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="2263d5aa-449b-448d-b567-cf24d9c15b74" containerName="extract-content" Oct 02 11:04:13 crc kubenswrapper[4934]: E1002 11:04:13.890155 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2263d5aa-449b-448d-b567-cf24d9c15b74" containerName="extract-utilities" Oct 02 11:04:13 crc kubenswrapper[4934]: I1002 11:04:13.890162 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="2263d5aa-449b-448d-b567-cf24d9c15b74" containerName="extract-utilities" Oct 02 11:04:13 crc kubenswrapper[4934]: I1002 11:04:13.890316 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="2263d5aa-449b-448d-b567-cf24d9c15b74" containerName="registry-server" Oct 02 11:04:13 crc kubenswrapper[4934]: I1002 11:04:13.891399 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hcscn" Oct 02 11:04:13 crc kubenswrapper[4934]: I1002 11:04:13.914230 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hcscn"] Oct 02 11:04:13 crc kubenswrapper[4934]: I1002 11:04:13.935233 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/304ced3f-1ad2-4d8c-a37c-6c652a40248c-catalog-content\") pod \"community-operators-hcscn\" (UID: \"304ced3f-1ad2-4d8c-a37c-6c652a40248c\") " pod="openshift-marketplace/community-operators-hcscn" Oct 02 11:04:13 crc kubenswrapper[4934]: I1002 11:04:13.935297 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/304ced3f-1ad2-4d8c-a37c-6c652a40248c-utilities\") pod \"community-operators-hcscn\" (UID: \"304ced3f-1ad2-4d8c-a37c-6c652a40248c\") " pod="openshift-marketplace/community-operators-hcscn" Oct 02 11:04:13 crc kubenswrapper[4934]: I1002 11:04:13.935343 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7mkd\" (UniqueName: \"kubernetes.io/projected/304ced3f-1ad2-4d8c-a37c-6c652a40248c-kube-api-access-v7mkd\") pod \"community-operators-hcscn\" (UID: \"304ced3f-1ad2-4d8c-a37c-6c652a40248c\") " pod="openshift-marketplace/community-operators-hcscn" Oct 02 11:04:14 crc kubenswrapper[4934]: I1002 11:04:14.036920 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/304ced3f-1ad2-4d8c-a37c-6c652a40248c-catalog-content\") pod \"community-operators-hcscn\" (UID: \"304ced3f-1ad2-4d8c-a37c-6c652a40248c\") " pod="openshift-marketplace/community-operators-hcscn" Oct 02 11:04:14 crc kubenswrapper[4934]: I1002 11:04:14.036987 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/304ced3f-1ad2-4d8c-a37c-6c652a40248c-utilities\") pod \"community-operators-hcscn\" (UID: \"304ced3f-1ad2-4d8c-a37c-6c652a40248c\") " pod="openshift-marketplace/community-operators-hcscn" Oct 02 11:04:14 crc kubenswrapper[4934]: I1002 11:04:14.037036 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7mkd\" (UniqueName: \"kubernetes.io/projected/304ced3f-1ad2-4d8c-a37c-6c652a40248c-kube-api-access-v7mkd\") pod \"community-operators-hcscn\" (UID: \"304ced3f-1ad2-4d8c-a37c-6c652a40248c\") " pod="openshift-marketplace/community-operators-hcscn" Oct 02 11:04:14 crc kubenswrapper[4934]: I1002 11:04:14.037905 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/304ced3f-1ad2-4d8c-a37c-6c652a40248c-catalog-content\") pod \"community-operators-hcscn\" (UID: \"304ced3f-1ad2-4d8c-a37c-6c652a40248c\") " pod="openshift-marketplace/community-operators-hcscn" Oct 02 11:04:14 crc kubenswrapper[4934]: I1002 11:04:14.038176 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/304ced3f-1ad2-4d8c-a37c-6c652a40248c-utilities\") pod \"community-operators-hcscn\" (UID: \"304ced3f-1ad2-4d8c-a37c-6c652a40248c\") " pod="openshift-marketplace/community-operators-hcscn" Oct 02 11:04:14 crc kubenswrapper[4934]: I1002 11:04:14.065538 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7mkd\" (UniqueName: \"kubernetes.io/projected/304ced3f-1ad2-4d8c-a37c-6c652a40248c-kube-api-access-v7mkd\") pod \"community-operators-hcscn\" (UID: \"304ced3f-1ad2-4d8c-a37c-6c652a40248c\") " pod="openshift-marketplace/community-operators-hcscn" Oct 02 11:04:14 crc kubenswrapper[4934]: I1002 11:04:14.212077 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hcscn" Oct 02 11:04:14 crc kubenswrapper[4934]: I1002 11:04:14.510362 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hcscn"] Oct 02 11:04:14 crc kubenswrapper[4934]: I1002 11:04:14.860449 4934 generic.go:334] "Generic (PLEG): container finished" podID="304ced3f-1ad2-4d8c-a37c-6c652a40248c" containerID="bfc9f8eee7f2ddda274c821a09ed19161b0ebf209a05065c5bcafffc7b240d6a" exitCode=0 Oct 02 11:04:14 crc kubenswrapper[4934]: I1002 11:04:14.860489 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hcscn" event={"ID":"304ced3f-1ad2-4d8c-a37c-6c652a40248c","Type":"ContainerDied","Data":"bfc9f8eee7f2ddda274c821a09ed19161b0ebf209a05065c5bcafffc7b240d6a"} Oct 02 11:04:14 crc kubenswrapper[4934]: I1002 11:04:14.860792 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hcscn" event={"ID":"304ced3f-1ad2-4d8c-a37c-6c652a40248c","Type":"ContainerStarted","Data":"49e922a026a8c783872c0e5aa79c5fc3483604841f9a4633bdb94a6a57866c85"} Oct 02 11:04:14 crc kubenswrapper[4934]: I1002 11:04:14.862484 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 11:04:16 crc kubenswrapper[4934]: I1002 11:04:16.877210 4934 generic.go:334] "Generic (PLEG): container finished" podID="304ced3f-1ad2-4d8c-a37c-6c652a40248c" containerID="1587c79cb21c7720298fb5ec6bb394ae226ed1363bc41b15721d4656f71180fe" exitCode=0 Oct 02 11:04:16 crc kubenswrapper[4934]: I1002 11:04:16.877262 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hcscn" event={"ID":"304ced3f-1ad2-4d8c-a37c-6c652a40248c","Type":"ContainerDied","Data":"1587c79cb21c7720298fb5ec6bb394ae226ed1363bc41b15721d4656f71180fe"} Oct 02 11:04:17 crc kubenswrapper[4934]: I1002 11:04:17.886382 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hcscn" event={"ID":"304ced3f-1ad2-4d8c-a37c-6c652a40248c","Type":"ContainerStarted","Data":"3fd6ed0d0b26d77614188ac26c023d865a57cbac5f64b23e072b32ad8d96cbc8"} Oct 02 11:04:17 crc kubenswrapper[4934]: I1002 11:04:17.904351 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hcscn" podStartSLOduration=2.426132393 podStartE2EDuration="4.904330738s" podCreationTimestamp="2025-10-02 11:04:13 +0000 UTC" firstStartedPulling="2025-10-02 11:04:14.862112661 +0000 UTC m=+4526.614754183" lastFinishedPulling="2025-10-02 11:04:17.340311006 +0000 UTC m=+4529.092952528" observedRunningTime="2025-10-02 11:04:17.903445803 +0000 UTC m=+4529.656087335" watchObservedRunningTime="2025-10-02 11:04:17.904330738 +0000 UTC m=+4529.656972280" Oct 02 11:04:22 crc kubenswrapper[4934]: I1002 11:04:22.913337 4934 scope.go:117] "RemoveContainer" containerID="ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39" Oct 02 11:04:22 crc kubenswrapper[4934]: E1002 11:04:22.914132 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:04:24 crc kubenswrapper[4934]: I1002 11:04:24.213069 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hcscn" Oct 02 11:04:24 crc kubenswrapper[4934]: I1002 11:04:24.213870 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hcscn" Oct 02 11:04:24 crc kubenswrapper[4934]: I1002 11:04:24.271678 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hcscn" Oct 02 11:04:24 crc kubenswrapper[4934]: I1002 11:04:24.987596 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hcscn" Oct 02 11:04:25 crc kubenswrapper[4934]: I1002 11:04:25.028877 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hcscn"] Oct 02 11:04:26 crc kubenswrapper[4934]: I1002 11:04:26.957532 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hcscn" podUID="304ced3f-1ad2-4d8c-a37c-6c652a40248c" containerName="registry-server" containerID="cri-o://3fd6ed0d0b26d77614188ac26c023d865a57cbac5f64b23e072b32ad8d96cbc8" gracePeriod=2 Oct 02 11:04:27 crc kubenswrapper[4934]: I1002 11:04:27.889108 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hcscn" Oct 02 11:04:27 crc kubenswrapper[4934]: I1002 11:04:27.945027 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/304ced3f-1ad2-4d8c-a37c-6c652a40248c-utilities\") pod \"304ced3f-1ad2-4d8c-a37c-6c652a40248c\" (UID: \"304ced3f-1ad2-4d8c-a37c-6c652a40248c\") " Oct 02 11:04:27 crc kubenswrapper[4934]: I1002 11:04:27.945176 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/304ced3f-1ad2-4d8c-a37c-6c652a40248c-catalog-content\") pod \"304ced3f-1ad2-4d8c-a37c-6c652a40248c\" (UID: \"304ced3f-1ad2-4d8c-a37c-6c652a40248c\") " Oct 02 11:04:27 crc kubenswrapper[4934]: I1002 11:04:27.945223 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7mkd\" (UniqueName: \"kubernetes.io/projected/304ced3f-1ad2-4d8c-a37c-6c652a40248c-kube-api-access-v7mkd\") pod \"304ced3f-1ad2-4d8c-a37c-6c652a40248c\" (UID: \"304ced3f-1ad2-4d8c-a37c-6c652a40248c\") " Oct 02 11:04:27 crc kubenswrapper[4934]: I1002 11:04:27.946096 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/304ced3f-1ad2-4d8c-a37c-6c652a40248c-utilities" (OuterVolumeSpecName: "utilities") pod "304ced3f-1ad2-4d8c-a37c-6c652a40248c" (UID: "304ced3f-1ad2-4d8c-a37c-6c652a40248c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:04:27 crc kubenswrapper[4934]: I1002 11:04:27.952668 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/304ced3f-1ad2-4d8c-a37c-6c652a40248c-kube-api-access-v7mkd" (OuterVolumeSpecName: "kube-api-access-v7mkd") pod "304ced3f-1ad2-4d8c-a37c-6c652a40248c" (UID: "304ced3f-1ad2-4d8c-a37c-6c652a40248c"). InnerVolumeSpecName "kube-api-access-v7mkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:04:27 crc kubenswrapper[4934]: I1002 11:04:27.972290 4934 generic.go:334] "Generic (PLEG): container finished" podID="304ced3f-1ad2-4d8c-a37c-6c652a40248c" containerID="3fd6ed0d0b26d77614188ac26c023d865a57cbac5f64b23e072b32ad8d96cbc8" exitCode=0 Oct 02 11:04:27 crc kubenswrapper[4934]: I1002 11:04:27.972339 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hcscn" event={"ID":"304ced3f-1ad2-4d8c-a37c-6c652a40248c","Type":"ContainerDied","Data":"3fd6ed0d0b26d77614188ac26c023d865a57cbac5f64b23e072b32ad8d96cbc8"} Oct 02 11:04:27 crc kubenswrapper[4934]: I1002 11:04:27.972364 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hcscn" event={"ID":"304ced3f-1ad2-4d8c-a37c-6c652a40248c","Type":"ContainerDied","Data":"49e922a026a8c783872c0e5aa79c5fc3483604841f9a4633bdb94a6a57866c85"} Oct 02 11:04:27 crc kubenswrapper[4934]: I1002 11:04:27.972380 4934 scope.go:117] "RemoveContainer" containerID="3fd6ed0d0b26d77614188ac26c023d865a57cbac5f64b23e072b32ad8d96cbc8" Oct 02 11:04:27 crc kubenswrapper[4934]: I1002 11:04:27.972496 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hcscn" Oct 02 11:04:27 crc kubenswrapper[4934]: I1002 11:04:27.995854 4934 scope.go:117] "RemoveContainer" containerID="1587c79cb21c7720298fb5ec6bb394ae226ed1363bc41b15721d4656f71180fe" Oct 02 11:04:28 crc kubenswrapper[4934]: I1002 11:04:28.003647 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/304ced3f-1ad2-4d8c-a37c-6c652a40248c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "304ced3f-1ad2-4d8c-a37c-6c652a40248c" (UID: "304ced3f-1ad2-4d8c-a37c-6c652a40248c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:04:28 crc kubenswrapper[4934]: I1002 11:04:28.010528 4934 scope.go:117] "RemoveContainer" containerID="bfc9f8eee7f2ddda274c821a09ed19161b0ebf209a05065c5bcafffc7b240d6a" Oct 02 11:04:28 crc kubenswrapper[4934]: I1002 11:04:28.036391 4934 scope.go:117] "RemoveContainer" containerID="3fd6ed0d0b26d77614188ac26c023d865a57cbac5f64b23e072b32ad8d96cbc8" Oct 02 11:04:28 crc kubenswrapper[4934]: E1002 11:04:28.037159 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fd6ed0d0b26d77614188ac26c023d865a57cbac5f64b23e072b32ad8d96cbc8\": container with ID starting with 3fd6ed0d0b26d77614188ac26c023d865a57cbac5f64b23e072b32ad8d96cbc8 not found: ID does not exist" containerID="3fd6ed0d0b26d77614188ac26c023d865a57cbac5f64b23e072b32ad8d96cbc8" Oct 02 11:04:28 crc kubenswrapper[4934]: I1002 11:04:28.037497 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fd6ed0d0b26d77614188ac26c023d865a57cbac5f64b23e072b32ad8d96cbc8"} err="failed to get container status \"3fd6ed0d0b26d77614188ac26c023d865a57cbac5f64b23e072b32ad8d96cbc8\": rpc error: code = NotFound desc = could not find container \"3fd6ed0d0b26d77614188ac26c023d865a57cbac5f64b23e072b32ad8d96cbc8\": container with ID starting with 3fd6ed0d0b26d77614188ac26c023d865a57cbac5f64b23e072b32ad8d96cbc8 not found: ID does not exist" Oct 02 11:04:28 crc kubenswrapper[4934]: I1002 11:04:28.037532 4934 scope.go:117] "RemoveContainer" containerID="1587c79cb21c7720298fb5ec6bb394ae226ed1363bc41b15721d4656f71180fe" Oct 02 11:04:28 crc kubenswrapper[4934]: E1002 11:04:28.038102 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1587c79cb21c7720298fb5ec6bb394ae226ed1363bc41b15721d4656f71180fe\": container with ID starting with 1587c79cb21c7720298fb5ec6bb394ae226ed1363bc41b15721d4656f71180fe not found: ID does not exist" containerID="1587c79cb21c7720298fb5ec6bb394ae226ed1363bc41b15721d4656f71180fe" Oct 02 11:04:28 crc kubenswrapper[4934]: I1002 11:04:28.038171 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1587c79cb21c7720298fb5ec6bb394ae226ed1363bc41b15721d4656f71180fe"} err="failed to get container status \"1587c79cb21c7720298fb5ec6bb394ae226ed1363bc41b15721d4656f71180fe\": rpc error: code = NotFound desc = could not find container \"1587c79cb21c7720298fb5ec6bb394ae226ed1363bc41b15721d4656f71180fe\": container with ID starting with 1587c79cb21c7720298fb5ec6bb394ae226ed1363bc41b15721d4656f71180fe not found: ID does not exist" Oct 02 11:04:28 crc kubenswrapper[4934]: I1002 11:04:28.038213 4934 scope.go:117] "RemoveContainer" containerID="bfc9f8eee7f2ddda274c821a09ed19161b0ebf209a05065c5bcafffc7b240d6a" Oct 02 11:04:28 crc kubenswrapper[4934]: E1002 11:04:28.038709 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfc9f8eee7f2ddda274c821a09ed19161b0ebf209a05065c5bcafffc7b240d6a\": container with ID starting with bfc9f8eee7f2ddda274c821a09ed19161b0ebf209a05065c5bcafffc7b240d6a not found: ID does not exist" containerID="bfc9f8eee7f2ddda274c821a09ed19161b0ebf209a05065c5bcafffc7b240d6a" Oct 02 11:04:28 crc kubenswrapper[4934]: I1002 11:04:28.038810 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfc9f8eee7f2ddda274c821a09ed19161b0ebf209a05065c5bcafffc7b240d6a"} err="failed to get container status \"bfc9f8eee7f2ddda274c821a09ed19161b0ebf209a05065c5bcafffc7b240d6a\": rpc error: code = NotFound desc = could not find container \"bfc9f8eee7f2ddda274c821a09ed19161b0ebf209a05065c5bcafffc7b240d6a\": container with ID starting with bfc9f8eee7f2ddda274c821a09ed19161b0ebf209a05065c5bcafffc7b240d6a not found: ID does not exist" Oct 02 11:04:28 crc kubenswrapper[4934]: I1002 11:04:28.046822 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/304ced3f-1ad2-4d8c-a37c-6c652a40248c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:04:28 crc kubenswrapper[4934]: I1002 11:04:28.046849 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7mkd\" (UniqueName: \"kubernetes.io/projected/304ced3f-1ad2-4d8c-a37c-6c652a40248c-kube-api-access-v7mkd\") on node \"crc\" DevicePath \"\"" Oct 02 11:04:28 crc kubenswrapper[4934]: I1002 11:04:28.046862 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/304ced3f-1ad2-4d8c-a37c-6c652a40248c-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:04:28 crc kubenswrapper[4934]: I1002 11:04:28.315979 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hcscn"] Oct 02 11:04:28 crc kubenswrapper[4934]: I1002 11:04:28.320502 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hcscn"] Oct 02 11:04:28 crc kubenswrapper[4934]: I1002 11:04:28.933962 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="304ced3f-1ad2-4d8c-a37c-6c652a40248c" path="/var/lib/kubelet/pods/304ced3f-1ad2-4d8c-a37c-6c652a40248c/volumes" Oct 02 11:04:35 crc kubenswrapper[4934]: I1002 11:04:35.913342 4934 scope.go:117] "RemoveContainer" containerID="ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39" Oct 02 11:04:35 crc kubenswrapper[4934]: E1002 11:04:35.913986 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:04:47 crc kubenswrapper[4934]: I1002 11:04:47.913788 4934 scope.go:117] "RemoveContainer" containerID="ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39" Oct 02 11:04:47 crc kubenswrapper[4934]: E1002 11:04:47.914664 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:04:58 crc kubenswrapper[4934]: I1002 11:04:58.918099 4934 scope.go:117] "RemoveContainer" containerID="ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39" Oct 02 11:04:58 crc kubenswrapper[4934]: E1002 11:04:58.919045 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:05:10 crc kubenswrapper[4934]: I1002 11:05:10.913181 4934 scope.go:117] "RemoveContainer" containerID="ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39" Oct 02 11:05:11 crc kubenswrapper[4934]: I1002 11:05:11.315119 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerStarted","Data":"71d39f1a175e0b8fa97659e0a291a0c09e00b393b1d0a0a2a372ae2e0143988b"} Oct 02 11:07:38 crc kubenswrapper[4934]: I1002 11:07:38.439278 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:07:38 crc kubenswrapper[4934]: I1002 11:07:38.439887 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:08:08 crc kubenswrapper[4934]: I1002 11:08:08.440020 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:08:08 crc kubenswrapper[4934]: I1002 11:08:08.440896 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:08:38 crc kubenswrapper[4934]: I1002 11:08:38.439889 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:08:38 crc kubenswrapper[4934]: I1002 11:08:38.440738 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:08:38 crc kubenswrapper[4934]: I1002 11:08:38.440798 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 11:08:38 crc kubenswrapper[4934]: I1002 11:08:38.441571 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"71d39f1a175e0b8fa97659e0a291a0c09e00b393b1d0a0a2a372ae2e0143988b"} pod="openshift-machine-config-operator/machine-config-daemon-djh5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 11:08:38 crc kubenswrapper[4934]: I1002 11:08:38.441716 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" containerID="cri-o://71d39f1a175e0b8fa97659e0a291a0c09e00b393b1d0a0a2a372ae2e0143988b" gracePeriod=600 Oct 02 11:08:38 crc kubenswrapper[4934]: I1002 11:08:38.961656 4934 generic.go:334] "Generic (PLEG): container finished" podID="71db06ef-05b0-4f58-b251-b27117a8500a" containerID="71d39f1a175e0b8fa97659e0a291a0c09e00b393b1d0a0a2a372ae2e0143988b" exitCode=0 Oct 02 11:08:38 crc kubenswrapper[4934]: I1002 11:08:38.961739 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerDied","Data":"71d39f1a175e0b8fa97659e0a291a0c09e00b393b1d0a0a2a372ae2e0143988b"} Oct 02 11:08:38 crc kubenswrapper[4934]: I1002 11:08:38.962040 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerStarted","Data":"1d6ec196f0e4bccca35b31484223dd4730674c6d770e39fd96fd7a619d6ba562"} Oct 02 11:08:38 crc kubenswrapper[4934]: I1002 11:08:38.962077 4934 scope.go:117] "RemoveContainer" containerID="ffb8701fb6846561f6d81fa6f17401b5336c5ebedfd38e752070bab217b4bf39" Oct 02 11:08:45 crc kubenswrapper[4934]: I1002 11:08:45.987177 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-htdcj"] Oct 02 11:08:45 crc kubenswrapper[4934]: E1002 11:08:45.988060 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="304ced3f-1ad2-4d8c-a37c-6c652a40248c" containerName="extract-utilities" Oct 02 11:08:45 crc kubenswrapper[4934]: I1002 11:08:45.988075 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="304ced3f-1ad2-4d8c-a37c-6c652a40248c" containerName="extract-utilities" Oct 02 11:08:45 crc kubenswrapper[4934]: E1002 11:08:45.988094 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="304ced3f-1ad2-4d8c-a37c-6c652a40248c" containerName="registry-server" Oct 02 11:08:45 crc kubenswrapper[4934]: I1002 11:08:45.988100 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="304ced3f-1ad2-4d8c-a37c-6c652a40248c" containerName="registry-server" Oct 02 11:08:45 crc kubenswrapper[4934]: E1002 11:08:45.988136 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="304ced3f-1ad2-4d8c-a37c-6c652a40248c" containerName="extract-content" Oct 02 11:08:45 crc kubenswrapper[4934]: I1002 11:08:45.988142 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="304ced3f-1ad2-4d8c-a37c-6c652a40248c" containerName="extract-content" Oct 02 11:08:45 crc kubenswrapper[4934]: I1002 11:08:45.988270 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="304ced3f-1ad2-4d8c-a37c-6c652a40248c" containerName="registry-server" Oct 02 11:08:45 crc kubenswrapper[4934]: I1002 11:08:45.989351 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-htdcj" Oct 02 11:08:45 crc kubenswrapper[4934]: I1002 11:08:45.995295 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-htdcj"] Oct 02 11:08:46 crc kubenswrapper[4934]: I1002 11:08:46.104471 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1a21af9-35d4-4c21-809c-524f2ddacecf-catalog-content\") pod \"redhat-operators-htdcj\" (UID: \"b1a21af9-35d4-4c21-809c-524f2ddacecf\") " pod="openshift-marketplace/redhat-operators-htdcj" Oct 02 11:08:46 crc kubenswrapper[4934]: I1002 11:08:46.104515 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzrkn\" (UniqueName: \"kubernetes.io/projected/b1a21af9-35d4-4c21-809c-524f2ddacecf-kube-api-access-fzrkn\") pod \"redhat-operators-htdcj\" (UID: \"b1a21af9-35d4-4c21-809c-524f2ddacecf\") " pod="openshift-marketplace/redhat-operators-htdcj" Oct 02 11:08:46 crc kubenswrapper[4934]: I1002 11:08:46.104628 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1a21af9-35d4-4c21-809c-524f2ddacecf-utilities\") pod \"redhat-operators-htdcj\" (UID: \"b1a21af9-35d4-4c21-809c-524f2ddacecf\") " pod="openshift-marketplace/redhat-operators-htdcj" Oct 02 11:08:46 crc kubenswrapper[4934]: I1002 11:08:46.205791 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1a21af9-35d4-4c21-809c-524f2ddacecf-utilities\") pod \"redhat-operators-htdcj\" (UID: \"b1a21af9-35d4-4c21-809c-524f2ddacecf\") " pod="openshift-marketplace/redhat-operators-htdcj" Oct 02 11:08:46 crc kubenswrapper[4934]: I1002 11:08:46.205905 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1a21af9-35d4-4c21-809c-524f2ddacecf-catalog-content\") pod \"redhat-operators-htdcj\" (UID: \"b1a21af9-35d4-4c21-809c-524f2ddacecf\") " pod="openshift-marketplace/redhat-operators-htdcj" Oct 02 11:08:46 crc kubenswrapper[4934]: I1002 11:08:46.205934 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzrkn\" (UniqueName: \"kubernetes.io/projected/b1a21af9-35d4-4c21-809c-524f2ddacecf-kube-api-access-fzrkn\") pod \"redhat-operators-htdcj\" (UID: \"b1a21af9-35d4-4c21-809c-524f2ddacecf\") " pod="openshift-marketplace/redhat-operators-htdcj" Oct 02 11:08:46 crc kubenswrapper[4934]: I1002 11:08:46.207775 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1a21af9-35d4-4c21-809c-524f2ddacecf-utilities\") pod \"redhat-operators-htdcj\" (UID: \"b1a21af9-35d4-4c21-809c-524f2ddacecf\") " pod="openshift-marketplace/redhat-operators-htdcj" Oct 02 11:08:46 crc kubenswrapper[4934]: I1002 11:08:46.209239 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1a21af9-35d4-4c21-809c-524f2ddacecf-catalog-content\") pod \"redhat-operators-htdcj\" (UID: \"b1a21af9-35d4-4c21-809c-524f2ddacecf\") " pod="openshift-marketplace/redhat-operators-htdcj" Oct 02 11:08:46 crc kubenswrapper[4934]: I1002 11:08:46.228418 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzrkn\" (UniqueName: \"kubernetes.io/projected/b1a21af9-35d4-4c21-809c-524f2ddacecf-kube-api-access-fzrkn\") pod \"redhat-operators-htdcj\" (UID: \"b1a21af9-35d4-4c21-809c-524f2ddacecf\") " pod="openshift-marketplace/redhat-operators-htdcj" Oct 02 11:08:46 crc kubenswrapper[4934]: I1002 11:08:46.322133 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-htdcj" Oct 02 11:08:46 crc kubenswrapper[4934]: I1002 11:08:46.744159 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-htdcj"] Oct 02 11:08:47 crc kubenswrapper[4934]: I1002 11:08:47.025059 4934 generic.go:334] "Generic (PLEG): container finished" podID="b1a21af9-35d4-4c21-809c-524f2ddacecf" containerID="1f1f9e3069af40baa1f38da86b405a26a04aca6408f60614ec0052a537c1c9a9" exitCode=0 Oct 02 11:08:47 crc kubenswrapper[4934]: I1002 11:08:47.025116 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-htdcj" event={"ID":"b1a21af9-35d4-4c21-809c-524f2ddacecf","Type":"ContainerDied","Data":"1f1f9e3069af40baa1f38da86b405a26a04aca6408f60614ec0052a537c1c9a9"} Oct 02 11:08:47 crc kubenswrapper[4934]: I1002 11:08:47.025345 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-htdcj" event={"ID":"b1a21af9-35d4-4c21-809c-524f2ddacecf","Type":"ContainerStarted","Data":"a8dcacfc6eb157aecf7093fde31b764ca6e468c934eecd203e16434ab8084e4a"} Oct 02 11:08:48 crc kubenswrapper[4934]: I1002 11:08:48.044452 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-htdcj" event={"ID":"b1a21af9-35d4-4c21-809c-524f2ddacecf","Type":"ContainerStarted","Data":"885f9734b95a379412a7581dc6015a9f73f37624fea2c46aed0245a1fc61985d"} Oct 02 11:08:49 crc kubenswrapper[4934]: I1002 11:08:49.055986 4934 generic.go:334] "Generic (PLEG): container finished" podID="b1a21af9-35d4-4c21-809c-524f2ddacecf" containerID="885f9734b95a379412a7581dc6015a9f73f37624fea2c46aed0245a1fc61985d" exitCode=0 Oct 02 11:08:49 crc kubenswrapper[4934]: I1002 11:08:49.056120 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-htdcj" event={"ID":"b1a21af9-35d4-4c21-809c-524f2ddacecf","Type":"ContainerDied","Data":"885f9734b95a379412a7581dc6015a9f73f37624fea2c46aed0245a1fc61985d"} Oct 02 11:08:50 crc kubenswrapper[4934]: I1002 11:08:50.067520 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-htdcj" event={"ID":"b1a21af9-35d4-4c21-809c-524f2ddacecf","Type":"ContainerStarted","Data":"b3118e6b900f9bc5470caf6ee269477548881230436f2da47bdc9a6d49269ef7"} Oct 02 11:08:50 crc kubenswrapper[4934]: I1002 11:08:50.092381 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-htdcj" podStartSLOduration=2.560882072 podStartE2EDuration="5.092361268s" podCreationTimestamp="2025-10-02 11:08:45 +0000 UTC" firstStartedPulling="2025-10-02 11:08:47.026244205 +0000 UTC m=+4798.778885727" lastFinishedPulling="2025-10-02 11:08:49.557723371 +0000 UTC m=+4801.310364923" observedRunningTime="2025-10-02 11:08:50.084528914 +0000 UTC m=+4801.837170496" watchObservedRunningTime="2025-10-02 11:08:50.092361268 +0000 UTC m=+4801.845002800" Oct 02 11:08:56 crc kubenswrapper[4934]: I1002 11:08:56.322812 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-htdcj" Oct 02 11:08:56 crc kubenswrapper[4934]: I1002 11:08:56.323382 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-htdcj" Oct 02 11:08:56 crc kubenswrapper[4934]: I1002 11:08:56.395506 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-htdcj" Oct 02 11:08:57 crc kubenswrapper[4934]: I1002 11:08:57.158952 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-htdcj" Oct 02 11:08:57 crc kubenswrapper[4934]: I1002 11:08:57.213489 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-htdcj"] Oct 02 11:08:59 crc kubenswrapper[4934]: I1002 11:08:59.137552 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-htdcj" podUID="b1a21af9-35d4-4c21-809c-524f2ddacecf" containerName="registry-server" containerID="cri-o://b3118e6b900f9bc5470caf6ee269477548881230436f2da47bdc9a6d49269ef7" gracePeriod=2 Oct 02 11:09:00 crc kubenswrapper[4934]: I1002 11:09:00.146771 4934 generic.go:334] "Generic (PLEG): container finished" podID="b1a21af9-35d4-4c21-809c-524f2ddacecf" containerID="b3118e6b900f9bc5470caf6ee269477548881230436f2da47bdc9a6d49269ef7" exitCode=0 Oct 02 11:09:00 crc kubenswrapper[4934]: I1002 11:09:00.146861 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-htdcj" event={"ID":"b1a21af9-35d4-4c21-809c-524f2ddacecf","Type":"ContainerDied","Data":"b3118e6b900f9bc5470caf6ee269477548881230436f2da47bdc9a6d49269ef7"} Oct 02 11:09:00 crc kubenswrapper[4934]: I1002 11:09:00.717814 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-htdcj" Oct 02 11:09:00 crc kubenswrapper[4934]: I1002 11:09:00.815048 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1a21af9-35d4-4c21-809c-524f2ddacecf-utilities\") pod \"b1a21af9-35d4-4c21-809c-524f2ddacecf\" (UID: \"b1a21af9-35d4-4c21-809c-524f2ddacecf\") " Oct 02 11:09:00 crc kubenswrapper[4934]: I1002 11:09:00.815127 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzrkn\" (UniqueName: \"kubernetes.io/projected/b1a21af9-35d4-4c21-809c-524f2ddacecf-kube-api-access-fzrkn\") pod \"b1a21af9-35d4-4c21-809c-524f2ddacecf\" (UID: \"b1a21af9-35d4-4c21-809c-524f2ddacecf\") " Oct 02 11:09:00 crc kubenswrapper[4934]: I1002 11:09:00.815187 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1a21af9-35d4-4c21-809c-524f2ddacecf-catalog-content\") pod \"b1a21af9-35d4-4c21-809c-524f2ddacecf\" (UID: \"b1a21af9-35d4-4c21-809c-524f2ddacecf\") " Oct 02 11:09:00 crc kubenswrapper[4934]: I1002 11:09:00.816054 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1a21af9-35d4-4c21-809c-524f2ddacecf-utilities" (OuterVolumeSpecName: "utilities") pod "b1a21af9-35d4-4c21-809c-524f2ddacecf" (UID: "b1a21af9-35d4-4c21-809c-524f2ddacecf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:09:00 crc kubenswrapper[4934]: I1002 11:09:00.820780 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1a21af9-35d4-4c21-809c-524f2ddacecf-kube-api-access-fzrkn" (OuterVolumeSpecName: "kube-api-access-fzrkn") pod "b1a21af9-35d4-4c21-809c-524f2ddacecf" (UID: "b1a21af9-35d4-4c21-809c-524f2ddacecf"). InnerVolumeSpecName "kube-api-access-fzrkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:09:00 crc kubenswrapper[4934]: I1002 11:09:00.917197 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzrkn\" (UniqueName: \"kubernetes.io/projected/b1a21af9-35d4-4c21-809c-524f2ddacecf-kube-api-access-fzrkn\") on node \"crc\" DevicePath \"\"" Oct 02 11:09:00 crc kubenswrapper[4934]: I1002 11:09:00.917262 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1a21af9-35d4-4c21-809c-524f2ddacecf-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:09:00 crc kubenswrapper[4934]: I1002 11:09:00.919395 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1a21af9-35d4-4c21-809c-524f2ddacecf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b1a21af9-35d4-4c21-809c-524f2ddacecf" (UID: "b1a21af9-35d4-4c21-809c-524f2ddacecf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:09:01 crc kubenswrapper[4934]: I1002 11:09:01.019562 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1a21af9-35d4-4c21-809c-524f2ddacecf-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:09:01 crc kubenswrapper[4934]: I1002 11:09:01.157904 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-htdcj" event={"ID":"b1a21af9-35d4-4c21-809c-524f2ddacecf","Type":"ContainerDied","Data":"a8dcacfc6eb157aecf7093fde31b764ca6e468c934eecd203e16434ab8084e4a"} Oct 02 11:09:01 crc kubenswrapper[4934]: I1002 11:09:01.158347 4934 scope.go:117] "RemoveContainer" containerID="b3118e6b900f9bc5470caf6ee269477548881230436f2da47bdc9a6d49269ef7" Oct 02 11:09:01 crc kubenswrapper[4934]: I1002 11:09:01.158111 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-htdcj" Oct 02 11:09:01 crc kubenswrapper[4934]: I1002 11:09:01.191324 4934 scope.go:117] "RemoveContainer" containerID="885f9734b95a379412a7581dc6015a9f73f37624fea2c46aed0245a1fc61985d" Oct 02 11:09:01 crc kubenswrapper[4934]: I1002 11:09:01.202199 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-htdcj"] Oct 02 11:09:01 crc kubenswrapper[4934]: I1002 11:09:01.211391 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-htdcj"] Oct 02 11:09:01 crc kubenswrapper[4934]: I1002 11:09:01.229421 4934 scope.go:117] "RemoveContainer" containerID="1f1f9e3069af40baa1f38da86b405a26a04aca6408f60614ec0052a537c1c9a9" Oct 02 11:09:02 crc kubenswrapper[4934]: I1002 11:09:02.924984 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1a21af9-35d4-4c21-809c-524f2ddacecf" path="/var/lib/kubelet/pods/b1a21af9-35d4-4c21-809c-524f2ddacecf/volumes" Oct 02 11:09:13 crc kubenswrapper[4934]: I1002 11:09:13.813685 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-k6wdn"] Oct 02 11:09:13 crc kubenswrapper[4934]: E1002 11:09:13.814842 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1a21af9-35d4-4c21-809c-524f2ddacecf" containerName="extract-content" Oct 02 11:09:13 crc kubenswrapper[4934]: I1002 11:09:13.814865 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1a21af9-35d4-4c21-809c-524f2ddacecf" containerName="extract-content" Oct 02 11:09:13 crc kubenswrapper[4934]: E1002 11:09:13.814894 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1a21af9-35d4-4c21-809c-524f2ddacecf" containerName="extract-utilities" Oct 02 11:09:13 crc kubenswrapper[4934]: I1002 11:09:13.814906 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1a21af9-35d4-4c21-809c-524f2ddacecf" containerName="extract-utilities" Oct 02 11:09:13 crc kubenswrapper[4934]: E1002 11:09:13.814926 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1a21af9-35d4-4c21-809c-524f2ddacecf" containerName="registry-server" Oct 02 11:09:13 crc kubenswrapper[4934]: I1002 11:09:13.814939 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1a21af9-35d4-4c21-809c-524f2ddacecf" containerName="registry-server" Oct 02 11:09:13 crc kubenswrapper[4934]: I1002 11:09:13.815197 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1a21af9-35d4-4c21-809c-524f2ddacecf" containerName="registry-server" Oct 02 11:09:13 crc kubenswrapper[4934]: I1002 11:09:13.817181 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k6wdn" Oct 02 11:09:13 crc kubenswrapper[4934]: I1002 11:09:13.823217 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k6wdn"] Oct 02 11:09:13 crc kubenswrapper[4934]: I1002 11:09:13.927377 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f43b4049-a7d7-4b8a-ae5a-ed903b844400-catalog-content\") pod \"certified-operators-k6wdn\" (UID: \"f43b4049-a7d7-4b8a-ae5a-ed903b844400\") " pod="openshift-marketplace/certified-operators-k6wdn" Oct 02 11:09:13 crc kubenswrapper[4934]: I1002 11:09:13.927483 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hxgb\" (UniqueName: \"kubernetes.io/projected/f43b4049-a7d7-4b8a-ae5a-ed903b844400-kube-api-access-4hxgb\") pod \"certified-operators-k6wdn\" (UID: \"f43b4049-a7d7-4b8a-ae5a-ed903b844400\") " pod="openshift-marketplace/certified-operators-k6wdn" Oct 02 11:09:13 crc kubenswrapper[4934]: I1002 11:09:13.927562 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f43b4049-a7d7-4b8a-ae5a-ed903b844400-utilities\") pod \"certified-operators-k6wdn\" (UID: \"f43b4049-a7d7-4b8a-ae5a-ed903b844400\") " pod="openshift-marketplace/certified-operators-k6wdn" Oct 02 11:09:14 crc kubenswrapper[4934]: I1002 11:09:14.028613 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f43b4049-a7d7-4b8a-ae5a-ed903b844400-utilities\") pod \"certified-operators-k6wdn\" (UID: \"f43b4049-a7d7-4b8a-ae5a-ed903b844400\") " pod="openshift-marketplace/certified-operators-k6wdn" Oct 02 11:09:14 crc kubenswrapper[4934]: I1002 11:09:14.028712 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f43b4049-a7d7-4b8a-ae5a-ed903b844400-catalog-content\") pod \"certified-operators-k6wdn\" (UID: \"f43b4049-a7d7-4b8a-ae5a-ed903b844400\") " pod="openshift-marketplace/certified-operators-k6wdn" Oct 02 11:09:14 crc kubenswrapper[4934]: I1002 11:09:14.029145 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f43b4049-a7d7-4b8a-ae5a-ed903b844400-utilities\") pod \"certified-operators-k6wdn\" (UID: \"f43b4049-a7d7-4b8a-ae5a-ed903b844400\") " pod="openshift-marketplace/certified-operators-k6wdn" Oct 02 11:09:14 crc kubenswrapper[4934]: I1002 11:09:14.029209 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f43b4049-a7d7-4b8a-ae5a-ed903b844400-catalog-content\") pod \"certified-operators-k6wdn\" (UID: \"f43b4049-a7d7-4b8a-ae5a-ed903b844400\") " pod="openshift-marketplace/certified-operators-k6wdn" Oct 02 11:09:14 crc kubenswrapper[4934]: I1002 11:09:14.029254 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hxgb\" (UniqueName: \"kubernetes.io/projected/f43b4049-a7d7-4b8a-ae5a-ed903b844400-kube-api-access-4hxgb\") pod \"certified-operators-k6wdn\" (UID: \"f43b4049-a7d7-4b8a-ae5a-ed903b844400\") " pod="openshift-marketplace/certified-operators-k6wdn" Oct 02 11:09:14 crc kubenswrapper[4934]: I1002 11:09:14.060735 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hxgb\" (UniqueName: \"kubernetes.io/projected/f43b4049-a7d7-4b8a-ae5a-ed903b844400-kube-api-access-4hxgb\") pod \"certified-operators-k6wdn\" (UID: \"f43b4049-a7d7-4b8a-ae5a-ed903b844400\") " pod="openshift-marketplace/certified-operators-k6wdn" Oct 02 11:09:14 crc kubenswrapper[4934]: I1002 11:09:14.158885 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k6wdn" Oct 02 11:09:14 crc kubenswrapper[4934]: I1002 11:09:14.437427 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k6wdn"] Oct 02 11:09:15 crc kubenswrapper[4934]: I1002 11:09:15.318262 4934 generic.go:334] "Generic (PLEG): container finished" podID="f43b4049-a7d7-4b8a-ae5a-ed903b844400" containerID="ed6277f9375e13d597dee258a194bb230b951b6d9ec3726589194864dd82a1ba" exitCode=0 Oct 02 11:09:15 crc kubenswrapper[4934]: I1002 11:09:15.318340 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k6wdn" event={"ID":"f43b4049-a7d7-4b8a-ae5a-ed903b844400","Type":"ContainerDied","Data":"ed6277f9375e13d597dee258a194bb230b951b6d9ec3726589194864dd82a1ba"} Oct 02 11:09:15 crc kubenswrapper[4934]: I1002 11:09:15.318448 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k6wdn" event={"ID":"f43b4049-a7d7-4b8a-ae5a-ed903b844400","Type":"ContainerStarted","Data":"57408dc8f550e1466078e2b84d13d1e71beba0a8ffd7414b0d61732cbaef213a"} Oct 02 11:09:15 crc kubenswrapper[4934]: I1002 11:09:15.321446 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 11:09:16 crc kubenswrapper[4934]: I1002 11:09:16.328776 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k6wdn" event={"ID":"f43b4049-a7d7-4b8a-ae5a-ed903b844400","Type":"ContainerStarted","Data":"14c7b71472caeb237555fc66e4b134d146a9ac8cd656511e2e01db4110404584"} Oct 02 11:09:17 crc kubenswrapper[4934]: I1002 11:09:17.341662 4934 generic.go:334] "Generic (PLEG): container finished" podID="f43b4049-a7d7-4b8a-ae5a-ed903b844400" containerID="14c7b71472caeb237555fc66e4b134d146a9ac8cd656511e2e01db4110404584" exitCode=0 Oct 02 11:09:17 crc kubenswrapper[4934]: I1002 11:09:17.341713 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k6wdn" event={"ID":"f43b4049-a7d7-4b8a-ae5a-ed903b844400","Type":"ContainerDied","Data":"14c7b71472caeb237555fc66e4b134d146a9ac8cd656511e2e01db4110404584"} Oct 02 11:09:18 crc kubenswrapper[4934]: I1002 11:09:18.351022 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k6wdn" event={"ID":"f43b4049-a7d7-4b8a-ae5a-ed903b844400","Type":"ContainerStarted","Data":"f546040b1ad3bb52fd76a8ec0d5610dd63131517e3256b440578f4e6633e7f90"} Oct 02 11:09:18 crc kubenswrapper[4934]: I1002 11:09:18.372833 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-k6wdn" podStartSLOduration=2.739785395 podStartE2EDuration="5.37281321s" podCreationTimestamp="2025-10-02 11:09:13 +0000 UTC" firstStartedPulling="2025-10-02 11:09:15.320985279 +0000 UTC m=+4827.073626841" lastFinishedPulling="2025-10-02 11:09:17.954013144 +0000 UTC m=+4829.706654656" observedRunningTime="2025-10-02 11:09:18.366496197 +0000 UTC m=+4830.119137729" watchObservedRunningTime="2025-10-02 11:09:18.37281321 +0000 UTC m=+4830.125454732" Oct 02 11:09:24 crc kubenswrapper[4934]: I1002 11:09:24.159447 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-k6wdn" Oct 02 11:09:24 crc kubenswrapper[4934]: I1002 11:09:24.159855 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-k6wdn" Oct 02 11:09:24 crc kubenswrapper[4934]: I1002 11:09:24.243287 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-k6wdn" Oct 02 11:09:24 crc kubenswrapper[4934]: I1002 11:09:24.435013 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-k6wdn" Oct 02 11:09:24 crc kubenswrapper[4934]: I1002 11:09:24.480045 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k6wdn"] Oct 02 11:09:26 crc kubenswrapper[4934]: I1002 11:09:26.413466 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-k6wdn" podUID="f43b4049-a7d7-4b8a-ae5a-ed903b844400" containerName="registry-server" containerID="cri-o://f546040b1ad3bb52fd76a8ec0d5610dd63131517e3256b440578f4e6633e7f90" gracePeriod=2 Oct 02 11:09:26 crc kubenswrapper[4934]: I1002 11:09:26.775187 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k6wdn" Oct 02 11:09:26 crc kubenswrapper[4934]: I1002 11:09:26.833973 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f43b4049-a7d7-4b8a-ae5a-ed903b844400-utilities\") pod \"f43b4049-a7d7-4b8a-ae5a-ed903b844400\" (UID: \"f43b4049-a7d7-4b8a-ae5a-ed903b844400\") " Oct 02 11:09:26 crc kubenswrapper[4934]: I1002 11:09:26.834049 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hxgb\" (UniqueName: \"kubernetes.io/projected/f43b4049-a7d7-4b8a-ae5a-ed903b844400-kube-api-access-4hxgb\") pod \"f43b4049-a7d7-4b8a-ae5a-ed903b844400\" (UID: \"f43b4049-a7d7-4b8a-ae5a-ed903b844400\") " Oct 02 11:09:26 crc kubenswrapper[4934]: I1002 11:09:26.834102 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f43b4049-a7d7-4b8a-ae5a-ed903b844400-catalog-content\") pod \"f43b4049-a7d7-4b8a-ae5a-ed903b844400\" (UID: \"f43b4049-a7d7-4b8a-ae5a-ed903b844400\") " Oct 02 11:09:26 crc kubenswrapper[4934]: I1002 11:09:26.835796 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f43b4049-a7d7-4b8a-ae5a-ed903b844400-utilities" (OuterVolumeSpecName: "utilities") pod "f43b4049-a7d7-4b8a-ae5a-ed903b844400" (UID: "f43b4049-a7d7-4b8a-ae5a-ed903b844400"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:09:26 crc kubenswrapper[4934]: I1002 11:09:26.840135 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f43b4049-a7d7-4b8a-ae5a-ed903b844400-kube-api-access-4hxgb" (OuterVolumeSpecName: "kube-api-access-4hxgb") pod "f43b4049-a7d7-4b8a-ae5a-ed903b844400" (UID: "f43b4049-a7d7-4b8a-ae5a-ed903b844400"). InnerVolumeSpecName "kube-api-access-4hxgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:09:26 crc kubenswrapper[4934]: I1002 11:09:26.889625 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f43b4049-a7d7-4b8a-ae5a-ed903b844400-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f43b4049-a7d7-4b8a-ae5a-ed903b844400" (UID: "f43b4049-a7d7-4b8a-ae5a-ed903b844400"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:09:26 crc kubenswrapper[4934]: I1002 11:09:26.936293 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hxgb\" (UniqueName: \"kubernetes.io/projected/f43b4049-a7d7-4b8a-ae5a-ed903b844400-kube-api-access-4hxgb\") on node \"crc\" DevicePath \"\"" Oct 02 11:09:26 crc kubenswrapper[4934]: I1002 11:09:26.936352 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f43b4049-a7d7-4b8a-ae5a-ed903b844400-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:09:26 crc kubenswrapper[4934]: I1002 11:09:26.936367 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f43b4049-a7d7-4b8a-ae5a-ed903b844400-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:09:27 crc kubenswrapper[4934]: I1002 11:09:27.423052 4934 generic.go:334] "Generic (PLEG): container finished" podID="f43b4049-a7d7-4b8a-ae5a-ed903b844400" containerID="f546040b1ad3bb52fd76a8ec0d5610dd63131517e3256b440578f4e6633e7f90" exitCode=0 Oct 02 11:09:27 crc kubenswrapper[4934]: I1002 11:09:27.423093 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k6wdn" event={"ID":"f43b4049-a7d7-4b8a-ae5a-ed903b844400","Type":"ContainerDied","Data":"f546040b1ad3bb52fd76a8ec0d5610dd63131517e3256b440578f4e6633e7f90"} Oct 02 11:09:27 crc kubenswrapper[4934]: I1002 11:09:27.423113 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k6wdn" Oct 02 11:09:27 crc kubenswrapper[4934]: I1002 11:09:27.423131 4934 scope.go:117] "RemoveContainer" containerID="f546040b1ad3bb52fd76a8ec0d5610dd63131517e3256b440578f4e6633e7f90" Oct 02 11:09:27 crc kubenswrapper[4934]: I1002 11:09:27.423120 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k6wdn" event={"ID":"f43b4049-a7d7-4b8a-ae5a-ed903b844400","Type":"ContainerDied","Data":"57408dc8f550e1466078e2b84d13d1e71beba0a8ffd7414b0d61732cbaef213a"} Oct 02 11:09:27 crc kubenswrapper[4934]: I1002 11:09:27.442925 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k6wdn"] Oct 02 11:09:27 crc kubenswrapper[4934]: I1002 11:09:27.448920 4934 scope.go:117] "RemoveContainer" containerID="14c7b71472caeb237555fc66e4b134d146a9ac8cd656511e2e01db4110404584" Oct 02 11:09:27 crc kubenswrapper[4934]: I1002 11:09:27.451118 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-k6wdn"] Oct 02 11:09:27 crc kubenswrapper[4934]: I1002 11:09:27.476853 4934 scope.go:117] "RemoveContainer" containerID="ed6277f9375e13d597dee258a194bb230b951b6d9ec3726589194864dd82a1ba" Oct 02 11:09:27 crc kubenswrapper[4934]: I1002 11:09:27.498906 4934 scope.go:117] "RemoveContainer" containerID="f546040b1ad3bb52fd76a8ec0d5610dd63131517e3256b440578f4e6633e7f90" Oct 02 11:09:27 crc kubenswrapper[4934]: E1002 11:09:27.499370 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f546040b1ad3bb52fd76a8ec0d5610dd63131517e3256b440578f4e6633e7f90\": container with ID starting with f546040b1ad3bb52fd76a8ec0d5610dd63131517e3256b440578f4e6633e7f90 not found: ID does not exist" containerID="f546040b1ad3bb52fd76a8ec0d5610dd63131517e3256b440578f4e6633e7f90" Oct 02 11:09:27 crc kubenswrapper[4934]: I1002 11:09:27.499410 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f546040b1ad3bb52fd76a8ec0d5610dd63131517e3256b440578f4e6633e7f90"} err="failed to get container status \"f546040b1ad3bb52fd76a8ec0d5610dd63131517e3256b440578f4e6633e7f90\": rpc error: code = NotFound desc = could not find container \"f546040b1ad3bb52fd76a8ec0d5610dd63131517e3256b440578f4e6633e7f90\": container with ID starting with f546040b1ad3bb52fd76a8ec0d5610dd63131517e3256b440578f4e6633e7f90 not found: ID does not exist" Oct 02 11:09:27 crc kubenswrapper[4934]: I1002 11:09:27.499440 4934 scope.go:117] "RemoveContainer" containerID="14c7b71472caeb237555fc66e4b134d146a9ac8cd656511e2e01db4110404584" Oct 02 11:09:27 crc kubenswrapper[4934]: E1002 11:09:27.499950 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14c7b71472caeb237555fc66e4b134d146a9ac8cd656511e2e01db4110404584\": container with ID starting with 14c7b71472caeb237555fc66e4b134d146a9ac8cd656511e2e01db4110404584 not found: ID does not exist" containerID="14c7b71472caeb237555fc66e4b134d146a9ac8cd656511e2e01db4110404584" Oct 02 11:09:27 crc kubenswrapper[4934]: I1002 11:09:27.499983 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14c7b71472caeb237555fc66e4b134d146a9ac8cd656511e2e01db4110404584"} err="failed to get container status \"14c7b71472caeb237555fc66e4b134d146a9ac8cd656511e2e01db4110404584\": rpc error: code = NotFound desc = could not find container \"14c7b71472caeb237555fc66e4b134d146a9ac8cd656511e2e01db4110404584\": container with ID starting with 14c7b71472caeb237555fc66e4b134d146a9ac8cd656511e2e01db4110404584 not found: ID does not exist" Oct 02 11:09:27 crc kubenswrapper[4934]: I1002 11:09:27.500006 4934 scope.go:117] "RemoveContainer" containerID="ed6277f9375e13d597dee258a194bb230b951b6d9ec3726589194864dd82a1ba" Oct 02 11:09:27 crc kubenswrapper[4934]: E1002 11:09:27.500261 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed6277f9375e13d597dee258a194bb230b951b6d9ec3726589194864dd82a1ba\": container with ID starting with ed6277f9375e13d597dee258a194bb230b951b6d9ec3726589194864dd82a1ba not found: ID does not exist" containerID="ed6277f9375e13d597dee258a194bb230b951b6d9ec3726589194864dd82a1ba" Oct 02 11:09:27 crc kubenswrapper[4934]: I1002 11:09:27.500299 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed6277f9375e13d597dee258a194bb230b951b6d9ec3726589194864dd82a1ba"} err="failed to get container status \"ed6277f9375e13d597dee258a194bb230b951b6d9ec3726589194864dd82a1ba\": rpc error: code = NotFound desc = could not find container \"ed6277f9375e13d597dee258a194bb230b951b6d9ec3726589194864dd82a1ba\": container with ID starting with ed6277f9375e13d597dee258a194bb230b951b6d9ec3726589194864dd82a1ba not found: ID does not exist" Oct 02 11:09:28 crc kubenswrapper[4934]: I1002 11:09:28.921182 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f43b4049-a7d7-4b8a-ae5a-ed903b844400" path="/var/lib/kubelet/pods/f43b4049-a7d7-4b8a-ae5a-ed903b844400/volumes" Oct 02 11:10:21 crc kubenswrapper[4934]: I1002 11:10:21.274598 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-x6r6h/must-gather-9zg6r"] Oct 02 11:10:21 crc kubenswrapper[4934]: E1002 11:10:21.275410 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f43b4049-a7d7-4b8a-ae5a-ed903b844400" containerName="extract-content" Oct 02 11:10:21 crc kubenswrapper[4934]: I1002 11:10:21.275425 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f43b4049-a7d7-4b8a-ae5a-ed903b844400" containerName="extract-content" Oct 02 11:10:21 crc kubenswrapper[4934]: E1002 11:10:21.275445 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f43b4049-a7d7-4b8a-ae5a-ed903b844400" containerName="extract-utilities" Oct 02 11:10:21 crc kubenswrapper[4934]: I1002 11:10:21.275454 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f43b4049-a7d7-4b8a-ae5a-ed903b844400" containerName="extract-utilities" Oct 02 11:10:21 crc kubenswrapper[4934]: E1002 11:10:21.275466 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f43b4049-a7d7-4b8a-ae5a-ed903b844400" containerName="registry-server" Oct 02 11:10:21 crc kubenswrapper[4934]: I1002 11:10:21.275478 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f43b4049-a7d7-4b8a-ae5a-ed903b844400" containerName="registry-server" Oct 02 11:10:21 crc kubenswrapper[4934]: I1002 11:10:21.275677 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f43b4049-a7d7-4b8a-ae5a-ed903b844400" containerName="registry-server" Oct 02 11:10:21 crc kubenswrapper[4934]: I1002 11:10:21.276693 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x6r6h/must-gather-9zg6r" Oct 02 11:10:21 crc kubenswrapper[4934]: I1002 11:10:21.283779 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-x6r6h"/"openshift-service-ca.crt" Oct 02 11:10:21 crc kubenswrapper[4934]: I1002 11:10:21.283779 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-x6r6h"/"kube-root-ca.crt" Oct 02 11:10:21 crc kubenswrapper[4934]: I1002 11:10:21.286909 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-x6r6h/must-gather-9zg6r"] Oct 02 11:10:21 crc kubenswrapper[4934]: I1002 11:10:21.312419 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcnz7\" (UniqueName: \"kubernetes.io/projected/d4c21786-bffe-40c4-9f2f-6f8046b5c1c0-kube-api-access-rcnz7\") pod \"must-gather-9zg6r\" (UID: \"d4c21786-bffe-40c4-9f2f-6f8046b5c1c0\") " pod="openshift-must-gather-x6r6h/must-gather-9zg6r" Oct 02 11:10:21 crc kubenswrapper[4934]: I1002 11:10:21.312490 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d4c21786-bffe-40c4-9f2f-6f8046b5c1c0-must-gather-output\") pod \"must-gather-9zg6r\" (UID: \"d4c21786-bffe-40c4-9f2f-6f8046b5c1c0\") " pod="openshift-must-gather-x6r6h/must-gather-9zg6r" Oct 02 11:10:21 crc kubenswrapper[4934]: I1002 11:10:21.416470 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d4c21786-bffe-40c4-9f2f-6f8046b5c1c0-must-gather-output\") pod \"must-gather-9zg6r\" (UID: \"d4c21786-bffe-40c4-9f2f-6f8046b5c1c0\") " pod="openshift-must-gather-x6r6h/must-gather-9zg6r" Oct 02 11:10:21 crc kubenswrapper[4934]: I1002 11:10:21.416561 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcnz7\" (UniqueName: \"kubernetes.io/projected/d4c21786-bffe-40c4-9f2f-6f8046b5c1c0-kube-api-access-rcnz7\") pod \"must-gather-9zg6r\" (UID: \"d4c21786-bffe-40c4-9f2f-6f8046b5c1c0\") " pod="openshift-must-gather-x6r6h/must-gather-9zg6r" Oct 02 11:10:21 crc kubenswrapper[4934]: I1002 11:10:21.417260 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d4c21786-bffe-40c4-9f2f-6f8046b5c1c0-must-gather-output\") pod \"must-gather-9zg6r\" (UID: \"d4c21786-bffe-40c4-9f2f-6f8046b5c1c0\") " pod="openshift-must-gather-x6r6h/must-gather-9zg6r" Oct 02 11:10:21 crc kubenswrapper[4934]: I1002 11:10:21.439300 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcnz7\" (UniqueName: \"kubernetes.io/projected/d4c21786-bffe-40c4-9f2f-6f8046b5c1c0-kube-api-access-rcnz7\") pod \"must-gather-9zg6r\" (UID: \"d4c21786-bffe-40c4-9f2f-6f8046b5c1c0\") " pod="openshift-must-gather-x6r6h/must-gather-9zg6r" Oct 02 11:10:21 crc kubenswrapper[4934]: I1002 11:10:21.594893 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x6r6h/must-gather-9zg6r" Oct 02 11:10:22 crc kubenswrapper[4934]: I1002 11:10:22.069225 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-x6r6h/must-gather-9zg6r"] Oct 02 11:10:22 crc kubenswrapper[4934]: I1002 11:10:22.828784 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x6r6h/must-gather-9zg6r" event={"ID":"d4c21786-bffe-40c4-9f2f-6f8046b5c1c0","Type":"ContainerStarted","Data":"3c68d07ab1948b92c2d54dc61c1bb9b613724b43311f08d3810199832afae749"} Oct 02 11:10:28 crc kubenswrapper[4934]: I1002 11:10:28.893935 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x6r6h/must-gather-9zg6r" event={"ID":"d4c21786-bffe-40c4-9f2f-6f8046b5c1c0","Type":"ContainerStarted","Data":"a5f816b66778a9f91b5c93f72af6a24afd969975cad255559a107fa3238e2671"} Oct 02 11:10:29 crc kubenswrapper[4934]: I1002 11:10:29.902152 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x6r6h/must-gather-9zg6r" event={"ID":"d4c21786-bffe-40c4-9f2f-6f8046b5c1c0","Type":"ContainerStarted","Data":"9e9176a751a90c013f39de73724866f30f1ec554ccfffdca838946ca95f745e9"} Oct 02 11:10:29 crc kubenswrapper[4934]: I1002 11:10:29.919960 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-x6r6h/must-gather-9zg6r" podStartSLOduration=2.518306021 podStartE2EDuration="8.919942109s" podCreationTimestamp="2025-10-02 11:10:21 +0000 UTC" firstStartedPulling="2025-10-02 11:10:22.079001886 +0000 UTC m=+4893.831643408" lastFinishedPulling="2025-10-02 11:10:28.480637974 +0000 UTC m=+4900.233279496" observedRunningTime="2025-10-02 11:10:29.917766659 +0000 UTC m=+4901.670408181" watchObservedRunningTime="2025-10-02 11:10:29.919942109 +0000 UTC m=+4901.672583631" Oct 02 11:10:38 crc kubenswrapper[4934]: I1002 11:10:38.439874 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:10:38 crc kubenswrapper[4934]: I1002 11:10:38.440462 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:11:08 crc kubenswrapper[4934]: I1002 11:11:08.439945 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:11:08 crc kubenswrapper[4934]: I1002 11:11:08.440455 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:11:18 crc kubenswrapper[4934]: I1002 11:11:18.704482 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5_94f3473a-92cb-428a-9cec-ad4de77d6a66/util/0.log" Oct 02 11:11:18 crc kubenswrapper[4934]: I1002 11:11:18.830934 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5_94f3473a-92cb-428a-9cec-ad4de77d6a66/util/0.log" Oct 02 11:11:18 crc kubenswrapper[4934]: I1002 11:11:18.872125 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5_94f3473a-92cb-428a-9cec-ad4de77d6a66/pull/0.log" Oct 02 11:11:18 crc kubenswrapper[4934]: I1002 11:11:18.872243 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5_94f3473a-92cb-428a-9cec-ad4de77d6a66/pull/0.log" Oct 02 11:11:19 crc kubenswrapper[4934]: I1002 11:11:19.001389 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5_94f3473a-92cb-428a-9cec-ad4de77d6a66/util/0.log" Oct 02 11:11:19 crc kubenswrapper[4934]: I1002 11:11:19.035088 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5_94f3473a-92cb-428a-9cec-ad4de77d6a66/extract/0.log" Oct 02 11:11:19 crc kubenswrapper[4934]: I1002 11:11:19.062849 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8966980b5e8cf8602d2f2556cdbe7c92b0c0dcd02d5e1c11c52fca56b994sp5_94f3473a-92cb-428a-9cec-ad4de77d6a66/pull/0.log" Oct 02 11:11:19 crc kubenswrapper[4934]: I1002 11:11:19.227370 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-f7f98cb69-lsrzt_2b00925c-c02d-4652-ae45-a3eadc31fda5/kube-rbac-proxy/0.log" Oct 02 11:11:19 crc kubenswrapper[4934]: I1002 11:11:19.234327 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-f7f98cb69-lsrzt_2b00925c-c02d-4652-ae45-a3eadc31fda5/manager/0.log" Oct 02 11:11:19 crc kubenswrapper[4934]: I1002 11:11:19.291897 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859cd486d-w2rxq_6a0e27a6-27cf-45bf-bb3d-adced5dd2fcd/kube-rbac-proxy/0.log" Oct 02 11:11:19 crc kubenswrapper[4934]: I1002 11:11:19.446998 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859cd486d-w2rxq_6a0e27a6-27cf-45bf-bb3d-adced5dd2fcd/manager/0.log" Oct 02 11:11:19 crc kubenswrapper[4934]: I1002 11:11:19.483709 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-77fb7bcf5b-jq9qb_5f887690-5ea6-401d-b37c-4623d5911e30/manager/0.log" Oct 02 11:11:19 crc kubenswrapper[4934]: I1002 11:11:19.507704 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-77fb7bcf5b-jq9qb_5f887690-5ea6-401d-b37c-4623d5911e30/kube-rbac-proxy/0.log" Oct 02 11:11:19 crc kubenswrapper[4934]: I1002 11:11:19.617823 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8bc4775b5-dh7nc_f315b606-7268-48c4-81b8-bb45bc0fb1a9/kube-rbac-proxy/0.log" Oct 02 11:11:19 crc kubenswrapper[4934]: I1002 11:11:19.711821 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8bc4775b5-dh7nc_f315b606-7268-48c4-81b8-bb45bc0fb1a9/manager/0.log" Oct 02 11:11:19 crc kubenswrapper[4934]: I1002 11:11:19.806158 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b4fc86755-xf5gv_da0df363-5daf-4162-8f99-93ebd8f9ee24/manager/0.log" Oct 02 11:11:19 crc kubenswrapper[4934]: I1002 11:11:19.825121 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b4fc86755-xf5gv_da0df363-5daf-4162-8f99-93ebd8f9ee24/kube-rbac-proxy/0.log" Oct 02 11:11:19 crc kubenswrapper[4934]: I1002 11:11:19.919790 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-679b4759bb-782sv_5d8d6294-fa14-4090-a472-862673713bf7/kube-rbac-proxy/0.log" Oct 02 11:11:19 crc kubenswrapper[4934]: I1002 11:11:19.993678 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-679b4759bb-782sv_5d8d6294-fa14-4090-a472-862673713bf7/manager/0.log" Oct 02 11:11:20 crc kubenswrapper[4934]: I1002 11:11:20.084556 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5c8fdc4d5c-6wzhw_ea2637c9-7013-4184-b7e6-3d429181b4aa/kube-rbac-proxy/0.log" Oct 02 11:11:20 crc kubenswrapper[4934]: I1002 11:11:20.198557 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5c8fdc4d5c-6wzhw_ea2637c9-7013-4184-b7e6-3d429181b4aa/manager/0.log" Oct 02 11:11:20 crc kubenswrapper[4934]: I1002 11:11:20.253002 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5f45cd594f-q7drf_d12830ea-c37d-46fe-952b-0d5019faf418/kube-rbac-proxy/0.log" Oct 02 11:11:20 crc kubenswrapper[4934]: I1002 11:11:20.288433 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5f45cd594f-q7drf_d12830ea-c37d-46fe-952b-0d5019faf418/manager/0.log" Oct 02 11:11:20 crc kubenswrapper[4934]: I1002 11:11:20.369155 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-59d7dc95cf-cswb5_1e12ae39-0145-42a5-ab78-855441de01fc/kube-rbac-proxy/0.log" Oct 02 11:11:20 crc kubenswrapper[4934]: I1002 11:11:20.468374 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-59d7dc95cf-cswb5_1e12ae39-0145-42a5-ab78-855441de01fc/manager/0.log" Oct 02 11:11:20 crc kubenswrapper[4934]: I1002 11:11:20.549651 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-b7cf8cb5f-sg4zk_0372f96c-947a-4625-8219-4dda67a14b3f/kube-rbac-proxy/0.log" Oct 02 11:11:20 crc kubenswrapper[4934]: I1002 11:11:20.569518 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-b7cf8cb5f-sg4zk_0372f96c-947a-4625-8219-4dda67a14b3f/manager/0.log" Oct 02 11:11:20 crc kubenswrapper[4934]: I1002 11:11:20.681054 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67bf5bb885-dxd5q_06a5a960-2aa6-4b1b-ae51-316873d545f6/kube-rbac-proxy/0.log" Oct 02 11:11:20 crc kubenswrapper[4934]: I1002 11:11:20.713812 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67bf5bb885-dxd5q_06a5a960-2aa6-4b1b-ae51-316873d545f6/manager/0.log" Oct 02 11:11:20 crc kubenswrapper[4934]: I1002 11:11:20.844237 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-54fbbfcd44-2dkpb_d411e947-e43e-4543-b2d7-177639563cc5/kube-rbac-proxy/0.log" Oct 02 11:11:20 crc kubenswrapper[4934]: I1002 11:11:20.916270 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-54fbbfcd44-2dkpb_d411e947-e43e-4543-b2d7-177639563cc5/manager/0.log" Oct 02 11:11:20 crc kubenswrapper[4934]: I1002 11:11:20.941605 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7fd5b6bbc6-7wntc_04d01fc6-fd8c-414d-b8b5-aab49056718d/kube-rbac-proxy/0.log" Oct 02 11:11:21 crc kubenswrapper[4934]: I1002 11:11:21.128175 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-75f8d67d86-dhf8d_5df0b2ac-a570-4594-8311-6c682eb07646/kube-rbac-proxy/0.log" Oct 02 11:11:21 crc kubenswrapper[4934]: I1002 11:11:21.144330 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7fd5b6bbc6-7wntc_04d01fc6-fd8c-414d-b8b5-aab49056718d/manager/0.log" Oct 02 11:11:21 crc kubenswrapper[4934]: I1002 11:11:21.179432 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-75f8d67d86-dhf8d_5df0b2ac-a570-4594-8311-6c682eb07646/manager/0.log" Oct 02 11:11:21 crc kubenswrapper[4934]: I1002 11:11:21.284758 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg_ce5aee9d-cb75-4b25-a795-75a362b717a3/kube-rbac-proxy/0.log" Oct 02 11:11:21 crc kubenswrapper[4934]: I1002 11:11:21.344962 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5f46788dd-jv7qg_ce5aee9d-cb75-4b25-a795-75a362b717a3/manager/0.log" Oct 02 11:11:21 crc kubenswrapper[4934]: I1002 11:11:21.456893 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-794f8d6c6d-ppdh4_f0e2fb9e-d1cb-42ab-b16b-b64f5e2240e2/kube-rbac-proxy/0.log" Oct 02 11:11:21 crc kubenswrapper[4934]: I1002 11:11:21.659002 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7c7c8d99ff-g2x7f_5d42fe89-99f8-436c-8401-835532319e14/kube-rbac-proxy/0.log" Oct 02 11:11:21 crc kubenswrapper[4934]: I1002 11:11:21.871957 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7c7c8d99ff-g2x7f_5d42fe89-99f8-436c-8401-835532319e14/operator/0.log" Oct 02 11:11:21 crc kubenswrapper[4934]: I1002 11:11:21.934285 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-fvkxx_c66499be-f2ac-4f3e-82e3-e90eed8922ce/registry-server/0.log" Oct 02 11:11:22 crc kubenswrapper[4934]: I1002 11:11:22.100924 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-84c745747f-qj6vc_a93ce93b-0718-4572-997c-3af703524717/kube-rbac-proxy/0.log" Oct 02 11:11:22 crc kubenswrapper[4934]: I1002 11:11:22.101355 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-794f8d6c6d-ppdh4_f0e2fb9e-d1cb-42ab-b16b-b64f5e2240e2/manager/0.log" Oct 02 11:11:22 crc kubenswrapper[4934]: I1002 11:11:22.144689 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-84c745747f-qj6vc_a93ce93b-0718-4572-997c-3af703524717/manager/0.log" Oct 02 11:11:22 crc kubenswrapper[4934]: I1002 11:11:22.190616 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-598c4c8547-6545m_0fd36c26-564f-4bda-b394-b21c674f5274/kube-rbac-proxy/0.log" Oct 02 11:11:22 crc kubenswrapper[4934]: I1002 11:11:22.295517 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-598c4c8547-6545m_0fd36c26-564f-4bda-b394-b21c674f5274/manager/0.log" Oct 02 11:11:22 crc kubenswrapper[4934]: I1002 11:11:22.300267 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-xbqhj_d5c07043-ed3a-4dd7-a685-f8dcc7bf578e/operator/0.log" Oct 02 11:11:22 crc kubenswrapper[4934]: I1002 11:11:22.382214 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-689b4f76c9-pk5ms_50cb1560-d7b8-4149-8fed-931e03814fb5/kube-rbac-proxy/0.log" Oct 02 11:11:22 crc kubenswrapper[4934]: I1002 11:11:22.468712 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-689b4f76c9-pk5ms_50cb1560-d7b8-4149-8fed-931e03814fb5/manager/0.log" Oct 02 11:11:22 crc kubenswrapper[4934]: I1002 11:11:22.499306 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-cb66d6b59-zxggl_dfe7cd03-27c4-4dc5-b875-ca762209e599/kube-rbac-proxy/0.log" Oct 02 11:11:22 crc kubenswrapper[4934]: I1002 11:11:22.616099 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-cb66d6b59-zxggl_dfe7cd03-27c4-4dc5-b875-ca762209e599/manager/0.log" Oct 02 11:11:22 crc kubenswrapper[4934]: I1002 11:11:22.654289 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-cbdf6dc66-xfxx5_0da92003-0ad0-4a25-9c55-16784a8fcc83/kube-rbac-proxy/0.log" Oct 02 11:11:22 crc kubenswrapper[4934]: I1002 11:11:22.677791 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-cbdf6dc66-xfxx5_0da92003-0ad0-4a25-9c55-16784a8fcc83/manager/0.log" Oct 02 11:11:22 crc kubenswrapper[4934]: I1002 11:11:22.760720 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-68d7bc5569-zsmk7_f470bad4-e97c-45f2-ac3f-3c3befc41094/kube-rbac-proxy/0.log" Oct 02 11:11:22 crc kubenswrapper[4934]: I1002 11:11:22.824809 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-68d7bc5569-zsmk7_f470bad4-e97c-45f2-ac3f-3c3befc41094/manager/0.log" Oct 02 11:11:36 crc kubenswrapper[4934]: I1002 11:11:36.255304 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-z2d6w_ef34af0b-654c-4519-9292-4e1a833b2ed3/control-plane-machine-set-operator/0.log" Oct 02 11:11:36 crc kubenswrapper[4934]: I1002 11:11:36.438311 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-t6dmx_8bbc993a-a1ea-48ae-a765-8f656d17329c/kube-rbac-proxy/0.log" Oct 02 11:11:36 crc kubenswrapper[4934]: I1002 11:11:36.487308 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-t6dmx_8bbc993a-a1ea-48ae-a765-8f656d17329c/machine-api-operator/0.log" Oct 02 11:11:38 crc kubenswrapper[4934]: I1002 11:11:38.439794 4934 patch_prober.go:28] interesting pod/machine-config-daemon-djh5z container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 11:11:38 crc kubenswrapper[4934]: I1002 11:11:38.440196 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 11:11:38 crc kubenswrapper[4934]: I1002 11:11:38.440251 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" Oct 02 11:11:38 crc kubenswrapper[4934]: I1002 11:11:38.441017 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1d6ec196f0e4bccca35b31484223dd4730674c6d770e39fd96fd7a619d6ba562"} pod="openshift-machine-config-operator/machine-config-daemon-djh5z" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 11:11:38 crc kubenswrapper[4934]: I1002 11:11:38.441096 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" containerName="machine-config-daemon" containerID="cri-o://1d6ec196f0e4bccca35b31484223dd4730674c6d770e39fd96fd7a619d6ba562" gracePeriod=600 Oct 02 11:11:39 crc kubenswrapper[4934]: E1002 11:11:39.196979 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:11:39 crc kubenswrapper[4934]: I1002 11:11:39.364470 4934 generic.go:334] "Generic (PLEG): container finished" podID="71db06ef-05b0-4f58-b251-b27117a8500a" containerID="1d6ec196f0e4bccca35b31484223dd4730674c6d770e39fd96fd7a619d6ba562" exitCode=0 Oct 02 11:11:39 crc kubenswrapper[4934]: I1002 11:11:39.364515 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" event={"ID":"71db06ef-05b0-4f58-b251-b27117a8500a","Type":"ContainerDied","Data":"1d6ec196f0e4bccca35b31484223dd4730674c6d770e39fd96fd7a619d6ba562"} Oct 02 11:11:39 crc kubenswrapper[4934]: I1002 11:11:39.364552 4934 scope.go:117] "RemoveContainer" containerID="71d39f1a175e0b8fa97659e0a291a0c09e00b393b1d0a0a2a372ae2e0143988b" Oct 02 11:11:39 crc kubenswrapper[4934]: I1002 11:11:39.365054 4934 scope.go:117] "RemoveContainer" containerID="1d6ec196f0e4bccca35b31484223dd4730674c6d770e39fd96fd7a619d6ba562" Oct 02 11:11:39 crc kubenswrapper[4934]: E1002 11:11:39.365374 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:11:47 crc kubenswrapper[4934]: I1002 11:11:47.290987 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-7d4cc89fcb-gp6dr_30f8f452-7988-4522-9506-f6de0a8d17d3/cert-manager-controller/0.log" Oct 02 11:11:47 crc kubenswrapper[4934]: I1002 11:11:47.474640 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7d9f95dbf-8w2z9_9fd4a62d-3068-4e62-8f90-7bf9c219ceee/cert-manager-cainjector/0.log" Oct 02 11:11:47 crc kubenswrapper[4934]: I1002 11:11:47.491780 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-d969966f-mfkpt_473f4679-38f6-49b2-8f0b-d874ab84678e/cert-manager-webhook/0.log" Oct 02 11:11:52 crc kubenswrapper[4934]: I1002 11:11:52.913044 4934 scope.go:117] "RemoveContainer" containerID="1d6ec196f0e4bccca35b31484223dd4730674c6d770e39fd96fd7a619d6ba562" Oct 02 11:11:52 crc kubenswrapper[4934]: E1002 11:11:52.913813 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:11:57 crc kubenswrapper[4934]: I1002 11:11:57.959996 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-msxrz_ef8e1eb9-668c-4542-9551-8a02b13222b9/nmstate-console-plugin/0.log" Oct 02 11:11:58 crc kubenswrapper[4934]: I1002 11:11:58.114022 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-mb8mr_687810e0-5141-42c9-a40b-5f04d045bd5d/nmstate-handler/0.log" Oct 02 11:11:58 crc kubenswrapper[4934]: I1002 11:11:58.168230 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-lwrzm_d705acd3-4167-430e-b33d-5966e485a9ce/kube-rbac-proxy/0.log" Oct 02 11:11:58 crc kubenswrapper[4934]: I1002 11:11:58.187414 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-lwrzm_d705acd3-4167-430e-b33d-5966e485a9ce/nmstate-metrics/0.log" Oct 02 11:11:58 crc kubenswrapper[4934]: I1002 11:11:58.313859 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-xcdzm_b9aa6395-2651-499c-afe7-ec95844b75d3/nmstate-operator/0.log" Oct 02 11:11:58 crc kubenswrapper[4934]: I1002 11:11:58.388097 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-w6z5d_f45d40d5-581b-47e0-a25e-9c550f92f689/nmstate-webhook/0.log" Oct 02 11:12:05 crc kubenswrapper[4934]: I1002 11:12:05.914063 4934 scope.go:117] "RemoveContainer" containerID="1d6ec196f0e4bccca35b31484223dd4730674c6d770e39fd96fd7a619d6ba562" Oct 02 11:12:05 crc kubenswrapper[4934]: E1002 11:12:05.914754 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:12:10 crc kubenswrapper[4934]: I1002 11:12:10.574748 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-74qhz_18786238-f18b-4b0a-b54d-e838c1092722/kube-rbac-proxy/0.log" Oct 02 11:12:10 crc kubenswrapper[4934]: I1002 11:12:10.772407 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jlmg7_853c1c50-69e1-47ca-9951-c75e52e11cc2/cp-frr-files/0.log" Oct 02 11:12:10 crc kubenswrapper[4934]: I1002 11:12:10.857041 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-74qhz_18786238-f18b-4b0a-b54d-e838c1092722/controller/0.log" Oct 02 11:12:10 crc kubenswrapper[4934]: I1002 11:12:10.943232 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jlmg7_853c1c50-69e1-47ca-9951-c75e52e11cc2/cp-frr-files/0.log" Oct 02 11:12:10 crc kubenswrapper[4934]: I1002 11:12:10.952127 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jlmg7_853c1c50-69e1-47ca-9951-c75e52e11cc2/cp-reloader/0.log" Oct 02 11:12:10 crc kubenswrapper[4934]: I1002 11:12:10.971267 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jlmg7_853c1c50-69e1-47ca-9951-c75e52e11cc2/cp-metrics/0.log" Oct 02 11:12:11 crc kubenswrapper[4934]: I1002 11:12:11.079925 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jlmg7_853c1c50-69e1-47ca-9951-c75e52e11cc2/cp-reloader/0.log" Oct 02 11:12:11 crc kubenswrapper[4934]: I1002 11:12:11.236622 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jlmg7_853c1c50-69e1-47ca-9951-c75e52e11cc2/cp-metrics/0.log" Oct 02 11:12:11 crc kubenswrapper[4934]: I1002 11:12:11.244315 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jlmg7_853c1c50-69e1-47ca-9951-c75e52e11cc2/cp-reloader/0.log" Oct 02 11:12:11 crc kubenswrapper[4934]: I1002 11:12:11.244463 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jlmg7_853c1c50-69e1-47ca-9951-c75e52e11cc2/cp-metrics/0.log" Oct 02 11:12:11 crc kubenswrapper[4934]: I1002 11:12:11.274123 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jlmg7_853c1c50-69e1-47ca-9951-c75e52e11cc2/cp-frr-files/0.log" Oct 02 11:12:11 crc kubenswrapper[4934]: I1002 11:12:11.482040 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jlmg7_853c1c50-69e1-47ca-9951-c75e52e11cc2/cp-reloader/0.log" Oct 02 11:12:11 crc kubenswrapper[4934]: I1002 11:12:11.516532 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jlmg7_853c1c50-69e1-47ca-9951-c75e52e11cc2/cp-frr-files/0.log" Oct 02 11:12:11 crc kubenswrapper[4934]: I1002 11:12:11.521256 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jlmg7_853c1c50-69e1-47ca-9951-c75e52e11cc2/controller/0.log" Oct 02 11:12:11 crc kubenswrapper[4934]: I1002 11:12:11.527847 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jlmg7_853c1c50-69e1-47ca-9951-c75e52e11cc2/cp-metrics/0.log" Oct 02 11:12:11 crc kubenswrapper[4934]: I1002 11:12:11.700839 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jlmg7_853c1c50-69e1-47ca-9951-c75e52e11cc2/kube-rbac-proxy/0.log" Oct 02 11:12:11 crc kubenswrapper[4934]: I1002 11:12:11.710941 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jlmg7_853c1c50-69e1-47ca-9951-c75e52e11cc2/frr-metrics/0.log" Oct 02 11:12:11 crc kubenswrapper[4934]: I1002 11:12:11.713624 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jlmg7_853c1c50-69e1-47ca-9951-c75e52e11cc2/kube-rbac-proxy-frr/0.log" Oct 02 11:12:11 crc kubenswrapper[4934]: I1002 11:12:11.915892 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jlmg7_853c1c50-69e1-47ca-9951-c75e52e11cc2/reloader/0.log" Oct 02 11:12:11 crc kubenswrapper[4934]: I1002 11:12:11.946449 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-pljrz_8cd364fe-896d-461f-b118-f757d17c3d1b/frr-k8s-webhook-server/0.log" Oct 02 11:12:12 crc kubenswrapper[4934]: I1002 11:12:12.163383 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-66d7dc4568-7b9dn_7fcd4706-2a88-4eda-9423-f535f059cbb3/manager/0.log" Oct 02 11:12:12 crc kubenswrapper[4934]: I1002 11:12:12.324546 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7b6d95f78d-q4xjj_b49e207c-8957-4b60-84f3-8b0b6c42579b/webhook-server/0.log" Oct 02 11:12:12 crc kubenswrapper[4934]: I1002 11:12:12.403499 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-m6ncq_8600909e-f2a1-4472-89ca-275d753fc807/kube-rbac-proxy/0.log" Oct 02 11:12:12 crc kubenswrapper[4934]: I1002 11:12:12.955557 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-m6ncq_8600909e-f2a1-4472-89ca-275d753fc807/speaker/0.log" Oct 02 11:12:13 crc kubenswrapper[4934]: I1002 11:12:13.004664 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jlmg7_853c1c50-69e1-47ca-9951-c75e52e11cc2/frr/0.log" Oct 02 11:12:16 crc kubenswrapper[4934]: I1002 11:12:16.913402 4934 scope.go:117] "RemoveContainer" containerID="1d6ec196f0e4bccca35b31484223dd4730674c6d770e39fd96fd7a619d6ba562" Oct 02 11:12:16 crc kubenswrapper[4934]: E1002 11:12:16.913867 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:12:23 crc kubenswrapper[4934]: I1002 11:12:23.441678 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw_537ea528-26ed-4970-a490-e625f7482003/util/0.log" Oct 02 11:12:23 crc kubenswrapper[4934]: I1002 11:12:23.596726 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw_537ea528-26ed-4970-a490-e625f7482003/util/0.log" Oct 02 11:12:23 crc kubenswrapper[4934]: I1002 11:12:23.633293 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw_537ea528-26ed-4970-a490-e625f7482003/pull/0.log" Oct 02 11:12:23 crc kubenswrapper[4934]: I1002 11:12:23.650426 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw_537ea528-26ed-4970-a490-e625f7482003/pull/0.log" Oct 02 11:12:23 crc kubenswrapper[4934]: I1002 11:12:23.876039 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw_537ea528-26ed-4970-a490-e625f7482003/pull/0.log" Oct 02 11:12:23 crc kubenswrapper[4934]: I1002 11:12:23.879722 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw_537ea528-26ed-4970-a490-e625f7482003/extract/0.log" Oct 02 11:12:23 crc kubenswrapper[4934]: I1002 11:12:23.920003 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69bmjpw_537ea528-26ed-4970-a490-e625f7482003/util/0.log" Oct 02 11:12:24 crc kubenswrapper[4934]: I1002 11:12:24.060533 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r_0107ccfd-b36f-4206-8f09-ff422dbc6a80/util/0.log" Oct 02 11:12:24 crc kubenswrapper[4934]: I1002 11:12:24.276854 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r_0107ccfd-b36f-4206-8f09-ff422dbc6a80/pull/0.log" Oct 02 11:12:24 crc kubenswrapper[4934]: I1002 11:12:24.300134 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r_0107ccfd-b36f-4206-8f09-ff422dbc6a80/pull/0.log" Oct 02 11:12:24 crc kubenswrapper[4934]: I1002 11:12:24.319108 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r_0107ccfd-b36f-4206-8f09-ff422dbc6a80/util/0.log" Oct 02 11:12:24 crc kubenswrapper[4934]: I1002 11:12:24.479666 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r_0107ccfd-b36f-4206-8f09-ff422dbc6a80/util/0.log" Oct 02 11:12:24 crc kubenswrapper[4934]: I1002 11:12:24.481456 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r_0107ccfd-b36f-4206-8f09-ff422dbc6a80/pull/0.log" Oct 02 11:12:24 crc kubenswrapper[4934]: I1002 11:12:24.515733 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2mbb9r_0107ccfd-b36f-4206-8f09-ff422dbc6a80/extract/0.log" Oct 02 11:12:24 crc kubenswrapper[4934]: I1002 11:12:24.652037 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-sfhmd_1df910b6-adca-48c6-8508-0e6535c84f6a/extract-utilities/0.log" Oct 02 11:12:24 crc kubenswrapper[4934]: I1002 11:12:24.801055 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-sfhmd_1df910b6-adca-48c6-8508-0e6535c84f6a/extract-content/0.log" Oct 02 11:12:24 crc kubenswrapper[4934]: I1002 11:12:24.846439 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-sfhmd_1df910b6-adca-48c6-8508-0e6535c84f6a/extract-content/0.log" Oct 02 11:12:24 crc kubenswrapper[4934]: I1002 11:12:24.856892 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-sfhmd_1df910b6-adca-48c6-8508-0e6535c84f6a/extract-utilities/0.log" Oct 02 11:12:25 crc kubenswrapper[4934]: I1002 11:12:25.018033 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-sfhmd_1df910b6-adca-48c6-8508-0e6535c84f6a/extract-utilities/0.log" Oct 02 11:12:25 crc kubenswrapper[4934]: I1002 11:12:25.048358 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-sfhmd_1df910b6-adca-48c6-8508-0e6535c84f6a/extract-content/0.log" Oct 02 11:12:25 crc kubenswrapper[4934]: I1002 11:12:25.198648 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zxq5v_31f0c5ea-f990-45d1-b501-53ce618fac45/extract-utilities/0.log" Oct 02 11:12:25 crc kubenswrapper[4934]: I1002 11:12:25.463981 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zxq5v_31f0c5ea-f990-45d1-b501-53ce618fac45/extract-content/0.log" Oct 02 11:12:25 crc kubenswrapper[4934]: I1002 11:12:25.519183 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zxq5v_31f0c5ea-f990-45d1-b501-53ce618fac45/extract-utilities/0.log" Oct 02 11:12:25 crc kubenswrapper[4934]: I1002 11:12:25.519186 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zxq5v_31f0c5ea-f990-45d1-b501-53ce618fac45/extract-content/0.log" Oct 02 11:12:25 crc kubenswrapper[4934]: I1002 11:12:25.656059 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-sfhmd_1df910b6-adca-48c6-8508-0e6535c84f6a/registry-server/0.log" Oct 02 11:12:25 crc kubenswrapper[4934]: I1002 11:12:25.807717 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zxq5v_31f0c5ea-f990-45d1-b501-53ce618fac45/extract-utilities/0.log" Oct 02 11:12:25 crc kubenswrapper[4934]: I1002 11:12:25.807717 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zxq5v_31f0c5ea-f990-45d1-b501-53ce618fac45/extract-content/0.log" Oct 02 11:12:26 crc kubenswrapper[4934]: I1002 11:12:26.018414 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl_89aa6f88-b397-4574-8ad9-6c855a2c8511/util/0.log" Oct 02 11:12:26 crc kubenswrapper[4934]: I1002 11:12:26.249761 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl_89aa6f88-b397-4574-8ad9-6c855a2c8511/util/0.log" Oct 02 11:12:26 crc kubenswrapper[4934]: I1002 11:12:26.297991 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl_89aa6f88-b397-4574-8ad9-6c855a2c8511/pull/0.log" Oct 02 11:12:26 crc kubenswrapper[4934]: I1002 11:12:26.301166 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl_89aa6f88-b397-4574-8ad9-6c855a2c8511/pull/0.log" Oct 02 11:12:26 crc kubenswrapper[4934]: I1002 11:12:26.418615 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zxq5v_31f0c5ea-f990-45d1-b501-53ce618fac45/registry-server/0.log" Oct 02 11:12:26 crc kubenswrapper[4934]: I1002 11:12:26.462596 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl_89aa6f88-b397-4574-8ad9-6c855a2c8511/pull/0.log" Oct 02 11:12:26 crc kubenswrapper[4934]: I1002 11:12:26.509117 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl_89aa6f88-b397-4574-8ad9-6c855a2c8511/util/0.log" Oct 02 11:12:26 crc kubenswrapper[4934]: I1002 11:12:26.515246 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cp5snl_89aa6f88-b397-4574-8ad9-6c855a2c8511/extract/0.log" Oct 02 11:12:26 crc kubenswrapper[4934]: I1002 11:12:26.642989 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-rhn6c_4f19193f-e9fe-4de1-a515-be021e499af8/marketplace-operator/0.log" Oct 02 11:12:26 crc kubenswrapper[4934]: I1002 11:12:26.700825 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4bvl5_b89bbfbf-3398-44a8-a54b-3fc5b55bec21/extract-utilities/0.log" Oct 02 11:12:26 crc kubenswrapper[4934]: I1002 11:12:26.889390 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4bvl5_b89bbfbf-3398-44a8-a54b-3fc5b55bec21/extract-content/0.log" Oct 02 11:12:26 crc kubenswrapper[4934]: I1002 11:12:26.910431 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4bvl5_b89bbfbf-3398-44a8-a54b-3fc5b55bec21/extract-content/0.log" Oct 02 11:12:26 crc kubenswrapper[4934]: I1002 11:12:26.938840 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4bvl5_b89bbfbf-3398-44a8-a54b-3fc5b55bec21/extract-utilities/0.log" Oct 02 11:12:27 crc kubenswrapper[4934]: I1002 11:12:27.090820 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4bvl5_b89bbfbf-3398-44a8-a54b-3fc5b55bec21/extract-utilities/0.log" Oct 02 11:12:27 crc kubenswrapper[4934]: I1002 11:12:27.092591 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4bvl5_b89bbfbf-3398-44a8-a54b-3fc5b55bec21/extract-content/0.log" Oct 02 11:12:27 crc kubenswrapper[4934]: I1002 11:12:27.213135 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vk4jw_599deecc-b859-4fc9-bb70-5675ef509325/extract-utilities/0.log" Oct 02 11:12:27 crc kubenswrapper[4934]: I1002 11:12:27.255889 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4bvl5_b89bbfbf-3398-44a8-a54b-3fc5b55bec21/registry-server/0.log" Oct 02 11:12:27 crc kubenswrapper[4934]: I1002 11:12:27.374315 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vk4jw_599deecc-b859-4fc9-bb70-5675ef509325/extract-utilities/0.log" Oct 02 11:12:27 crc kubenswrapper[4934]: I1002 11:12:27.406615 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vk4jw_599deecc-b859-4fc9-bb70-5675ef509325/extract-content/0.log" Oct 02 11:12:27 crc kubenswrapper[4934]: I1002 11:12:27.415260 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vk4jw_599deecc-b859-4fc9-bb70-5675ef509325/extract-content/0.log" Oct 02 11:12:27 crc kubenswrapper[4934]: I1002 11:12:27.578099 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vk4jw_599deecc-b859-4fc9-bb70-5675ef509325/extract-content/0.log" Oct 02 11:12:27 crc kubenswrapper[4934]: I1002 11:12:27.597293 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vk4jw_599deecc-b859-4fc9-bb70-5675ef509325/extract-utilities/0.log" Oct 02 11:12:28 crc kubenswrapper[4934]: I1002 11:12:28.212808 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vk4jw_599deecc-b859-4fc9-bb70-5675ef509325/registry-server/0.log" Oct 02 11:12:28 crc kubenswrapper[4934]: I1002 11:12:28.919122 4934 scope.go:117] "RemoveContainer" containerID="1d6ec196f0e4bccca35b31484223dd4730674c6d770e39fd96fd7a619d6ba562" Oct 02 11:12:28 crc kubenswrapper[4934]: E1002 11:12:28.919708 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:12:41 crc kubenswrapper[4934]: I1002 11:12:41.914937 4934 scope.go:117] "RemoveContainer" containerID="1d6ec196f0e4bccca35b31484223dd4730674c6d770e39fd96fd7a619d6ba562" Oct 02 11:12:41 crc kubenswrapper[4934]: E1002 11:12:41.915819 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:12:56 crc kubenswrapper[4934]: I1002 11:12:56.913776 4934 scope.go:117] "RemoveContainer" containerID="1d6ec196f0e4bccca35b31484223dd4730674c6d770e39fd96fd7a619d6ba562" Oct 02 11:12:56 crc kubenswrapper[4934]: E1002 11:12:56.914748 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:13:04 crc kubenswrapper[4934]: I1002 11:13:04.928984 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-v9rdc"] Oct 02 11:13:04 crc kubenswrapper[4934]: I1002 11:13:04.931441 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v9rdc" Oct 02 11:13:04 crc kubenswrapper[4934]: I1002 11:13:04.941585 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v9rdc"] Oct 02 11:13:05 crc kubenswrapper[4934]: I1002 11:13:05.060679 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nws2r\" (UniqueName: \"kubernetes.io/projected/b4365e38-5e5f-447f-a92c-d8b9870e6b31-kube-api-access-nws2r\") pod \"redhat-marketplace-v9rdc\" (UID: \"b4365e38-5e5f-447f-a92c-d8b9870e6b31\") " pod="openshift-marketplace/redhat-marketplace-v9rdc" Oct 02 11:13:05 crc kubenswrapper[4934]: I1002 11:13:05.060747 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4365e38-5e5f-447f-a92c-d8b9870e6b31-catalog-content\") pod \"redhat-marketplace-v9rdc\" (UID: \"b4365e38-5e5f-447f-a92c-d8b9870e6b31\") " pod="openshift-marketplace/redhat-marketplace-v9rdc" Oct 02 11:13:05 crc kubenswrapper[4934]: I1002 11:13:05.060792 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4365e38-5e5f-447f-a92c-d8b9870e6b31-utilities\") pod \"redhat-marketplace-v9rdc\" (UID: \"b4365e38-5e5f-447f-a92c-d8b9870e6b31\") " pod="openshift-marketplace/redhat-marketplace-v9rdc" Oct 02 11:13:05 crc kubenswrapper[4934]: I1002 11:13:05.162320 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4365e38-5e5f-447f-a92c-d8b9870e6b31-utilities\") pod \"redhat-marketplace-v9rdc\" (UID: \"b4365e38-5e5f-447f-a92c-d8b9870e6b31\") " pod="openshift-marketplace/redhat-marketplace-v9rdc" Oct 02 11:13:05 crc kubenswrapper[4934]: I1002 11:13:05.162436 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nws2r\" (UniqueName: \"kubernetes.io/projected/b4365e38-5e5f-447f-a92c-d8b9870e6b31-kube-api-access-nws2r\") pod \"redhat-marketplace-v9rdc\" (UID: \"b4365e38-5e5f-447f-a92c-d8b9870e6b31\") " pod="openshift-marketplace/redhat-marketplace-v9rdc" Oct 02 11:13:05 crc kubenswrapper[4934]: I1002 11:13:05.162467 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4365e38-5e5f-447f-a92c-d8b9870e6b31-catalog-content\") pod \"redhat-marketplace-v9rdc\" (UID: \"b4365e38-5e5f-447f-a92c-d8b9870e6b31\") " pod="openshift-marketplace/redhat-marketplace-v9rdc" Oct 02 11:13:05 crc kubenswrapper[4934]: I1002 11:13:05.163000 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4365e38-5e5f-447f-a92c-d8b9870e6b31-utilities\") pod \"redhat-marketplace-v9rdc\" (UID: \"b4365e38-5e5f-447f-a92c-d8b9870e6b31\") " pod="openshift-marketplace/redhat-marketplace-v9rdc" Oct 02 11:13:05 crc kubenswrapper[4934]: I1002 11:13:05.163067 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4365e38-5e5f-447f-a92c-d8b9870e6b31-catalog-content\") pod \"redhat-marketplace-v9rdc\" (UID: \"b4365e38-5e5f-447f-a92c-d8b9870e6b31\") " pod="openshift-marketplace/redhat-marketplace-v9rdc" Oct 02 11:13:05 crc kubenswrapper[4934]: I1002 11:13:05.190782 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nws2r\" (UniqueName: \"kubernetes.io/projected/b4365e38-5e5f-447f-a92c-d8b9870e6b31-kube-api-access-nws2r\") pod \"redhat-marketplace-v9rdc\" (UID: \"b4365e38-5e5f-447f-a92c-d8b9870e6b31\") " pod="openshift-marketplace/redhat-marketplace-v9rdc" Oct 02 11:13:05 crc kubenswrapper[4934]: I1002 11:13:05.262130 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v9rdc" Oct 02 11:13:05 crc kubenswrapper[4934]: I1002 11:13:05.660823 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v9rdc"] Oct 02 11:13:05 crc kubenswrapper[4934]: I1002 11:13:05.914025 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9rdc" event={"ID":"b4365e38-5e5f-447f-a92c-d8b9870e6b31","Type":"ContainerStarted","Data":"3d3661d4335cbf1df0b1c6fb6c20ac22ca6e8058b029545a46e86e908eb8dab8"} Oct 02 11:13:06 crc kubenswrapper[4934]: I1002 11:13:06.921852 4934 generic.go:334] "Generic (PLEG): container finished" podID="b4365e38-5e5f-447f-a92c-d8b9870e6b31" containerID="adbf17d31317d364841f139cc80038389e30616329d51dabed6cf4d3cdebb4fd" exitCode=0 Oct 02 11:13:06 crc kubenswrapper[4934]: I1002 11:13:06.923086 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9rdc" event={"ID":"b4365e38-5e5f-447f-a92c-d8b9870e6b31","Type":"ContainerDied","Data":"adbf17d31317d364841f139cc80038389e30616329d51dabed6cf4d3cdebb4fd"} Oct 02 11:13:08 crc kubenswrapper[4934]: I1002 11:13:08.936546 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9rdc" event={"ID":"b4365e38-5e5f-447f-a92c-d8b9870e6b31","Type":"ContainerDied","Data":"dc55b5e2b06913bd776cc3cb499264b480ba13983dcfec8245b6b125888bda1e"} Oct 02 11:13:08 crc kubenswrapper[4934]: I1002 11:13:08.936382 4934 generic.go:334] "Generic (PLEG): container finished" podID="b4365e38-5e5f-447f-a92c-d8b9870e6b31" containerID="dc55b5e2b06913bd776cc3cb499264b480ba13983dcfec8245b6b125888bda1e" exitCode=0 Oct 02 11:13:09 crc kubenswrapper[4934]: I1002 11:13:09.914526 4934 scope.go:117] "RemoveContainer" containerID="1d6ec196f0e4bccca35b31484223dd4730674c6d770e39fd96fd7a619d6ba562" Oct 02 11:13:09 crc kubenswrapper[4934]: E1002 11:13:09.915388 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:13:09 crc kubenswrapper[4934]: I1002 11:13:09.949643 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9rdc" event={"ID":"b4365e38-5e5f-447f-a92c-d8b9870e6b31","Type":"ContainerStarted","Data":"5e3ea7b352e92ac01436df12b034e1b1c0e5c66742216f79fd703c305a9b2a3f"} Oct 02 11:13:09 crc kubenswrapper[4934]: I1002 11:13:09.973888 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-v9rdc" podStartSLOduration=3.354093996 podStartE2EDuration="5.973865899s" podCreationTimestamp="2025-10-02 11:13:04 +0000 UTC" firstStartedPulling="2025-10-02 11:13:06.92356977 +0000 UTC m=+5058.676211292" lastFinishedPulling="2025-10-02 11:13:09.543341673 +0000 UTC m=+5061.295983195" observedRunningTime="2025-10-02 11:13:09.970304762 +0000 UTC m=+5061.722946334" watchObservedRunningTime="2025-10-02 11:13:09.973865899 +0000 UTC m=+5061.726507441" Oct 02 11:13:15 crc kubenswrapper[4934]: I1002 11:13:15.263510 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-v9rdc" Oct 02 11:13:15 crc kubenswrapper[4934]: I1002 11:13:15.264014 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-v9rdc" Oct 02 11:13:15 crc kubenswrapper[4934]: I1002 11:13:15.306118 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-v9rdc" Oct 02 11:13:16 crc kubenswrapper[4934]: I1002 11:13:16.044985 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-v9rdc" Oct 02 11:13:16 crc kubenswrapper[4934]: I1002 11:13:16.094885 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v9rdc"] Oct 02 11:13:18 crc kubenswrapper[4934]: I1002 11:13:18.020365 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-v9rdc" podUID="b4365e38-5e5f-447f-a92c-d8b9870e6b31" containerName="registry-server" containerID="cri-o://5e3ea7b352e92ac01436df12b034e1b1c0e5c66742216f79fd703c305a9b2a3f" gracePeriod=2 Oct 02 11:13:18 crc kubenswrapper[4934]: I1002 11:13:18.478198 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v9rdc" Oct 02 11:13:18 crc kubenswrapper[4934]: I1002 11:13:18.570035 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4365e38-5e5f-447f-a92c-d8b9870e6b31-catalog-content\") pod \"b4365e38-5e5f-447f-a92c-d8b9870e6b31\" (UID: \"b4365e38-5e5f-447f-a92c-d8b9870e6b31\") " Oct 02 11:13:18 crc kubenswrapper[4934]: I1002 11:13:18.570216 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4365e38-5e5f-447f-a92c-d8b9870e6b31-utilities\") pod \"b4365e38-5e5f-447f-a92c-d8b9870e6b31\" (UID: \"b4365e38-5e5f-447f-a92c-d8b9870e6b31\") " Oct 02 11:13:18 crc kubenswrapper[4934]: I1002 11:13:18.570288 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nws2r\" (UniqueName: \"kubernetes.io/projected/b4365e38-5e5f-447f-a92c-d8b9870e6b31-kube-api-access-nws2r\") pod \"b4365e38-5e5f-447f-a92c-d8b9870e6b31\" (UID: \"b4365e38-5e5f-447f-a92c-d8b9870e6b31\") " Oct 02 11:13:18 crc kubenswrapper[4934]: I1002 11:13:18.571618 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4365e38-5e5f-447f-a92c-d8b9870e6b31-utilities" (OuterVolumeSpecName: "utilities") pod "b4365e38-5e5f-447f-a92c-d8b9870e6b31" (UID: "b4365e38-5e5f-447f-a92c-d8b9870e6b31"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:13:18 crc kubenswrapper[4934]: I1002 11:13:18.577140 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4365e38-5e5f-447f-a92c-d8b9870e6b31-kube-api-access-nws2r" (OuterVolumeSpecName: "kube-api-access-nws2r") pod "b4365e38-5e5f-447f-a92c-d8b9870e6b31" (UID: "b4365e38-5e5f-447f-a92c-d8b9870e6b31"). InnerVolumeSpecName "kube-api-access-nws2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:13:18 crc kubenswrapper[4934]: I1002 11:13:18.583393 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4365e38-5e5f-447f-a92c-d8b9870e6b31-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b4365e38-5e5f-447f-a92c-d8b9870e6b31" (UID: "b4365e38-5e5f-447f-a92c-d8b9870e6b31"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:13:18 crc kubenswrapper[4934]: I1002 11:13:18.672014 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4365e38-5e5f-447f-a92c-d8b9870e6b31-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:13:18 crc kubenswrapper[4934]: I1002 11:13:18.672065 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4365e38-5e5f-447f-a92c-d8b9870e6b31-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:13:18 crc kubenswrapper[4934]: I1002 11:13:18.672081 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nws2r\" (UniqueName: \"kubernetes.io/projected/b4365e38-5e5f-447f-a92c-d8b9870e6b31-kube-api-access-nws2r\") on node \"crc\" DevicePath \"\"" Oct 02 11:13:19 crc kubenswrapper[4934]: I1002 11:13:19.036586 4934 generic.go:334] "Generic (PLEG): container finished" podID="b4365e38-5e5f-447f-a92c-d8b9870e6b31" containerID="5e3ea7b352e92ac01436df12b034e1b1c0e5c66742216f79fd703c305a9b2a3f" exitCode=0 Oct 02 11:13:19 crc kubenswrapper[4934]: I1002 11:13:19.036694 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9rdc" event={"ID":"b4365e38-5e5f-447f-a92c-d8b9870e6b31","Type":"ContainerDied","Data":"5e3ea7b352e92ac01436df12b034e1b1c0e5c66742216f79fd703c305a9b2a3f"} Oct 02 11:13:19 crc kubenswrapper[4934]: I1002 11:13:19.036752 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9rdc" event={"ID":"b4365e38-5e5f-447f-a92c-d8b9870e6b31","Type":"ContainerDied","Data":"3d3661d4335cbf1df0b1c6fb6c20ac22ca6e8058b029545a46e86e908eb8dab8"} Oct 02 11:13:19 crc kubenswrapper[4934]: I1002 11:13:19.036788 4934 scope.go:117] "RemoveContainer" containerID="5e3ea7b352e92ac01436df12b034e1b1c0e5c66742216f79fd703c305a9b2a3f" Oct 02 11:13:19 crc kubenswrapper[4934]: I1002 11:13:19.036837 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v9rdc" Oct 02 11:13:19 crc kubenswrapper[4934]: I1002 11:13:19.062039 4934 scope.go:117] "RemoveContainer" containerID="dc55b5e2b06913bd776cc3cb499264b480ba13983dcfec8245b6b125888bda1e" Oct 02 11:13:19 crc kubenswrapper[4934]: I1002 11:13:19.071709 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v9rdc"] Oct 02 11:13:19 crc kubenswrapper[4934]: I1002 11:13:19.079619 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-v9rdc"] Oct 02 11:13:19 crc kubenswrapper[4934]: I1002 11:13:19.086831 4934 scope.go:117] "RemoveContainer" containerID="adbf17d31317d364841f139cc80038389e30616329d51dabed6cf4d3cdebb4fd" Oct 02 11:13:19 crc kubenswrapper[4934]: I1002 11:13:19.112747 4934 scope.go:117] "RemoveContainer" containerID="5e3ea7b352e92ac01436df12b034e1b1c0e5c66742216f79fd703c305a9b2a3f" Oct 02 11:13:19 crc kubenswrapper[4934]: E1002 11:13:19.113227 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e3ea7b352e92ac01436df12b034e1b1c0e5c66742216f79fd703c305a9b2a3f\": container with ID starting with 5e3ea7b352e92ac01436df12b034e1b1c0e5c66742216f79fd703c305a9b2a3f not found: ID does not exist" containerID="5e3ea7b352e92ac01436df12b034e1b1c0e5c66742216f79fd703c305a9b2a3f" Oct 02 11:13:19 crc kubenswrapper[4934]: I1002 11:13:19.113255 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e3ea7b352e92ac01436df12b034e1b1c0e5c66742216f79fd703c305a9b2a3f"} err="failed to get container status \"5e3ea7b352e92ac01436df12b034e1b1c0e5c66742216f79fd703c305a9b2a3f\": rpc error: code = NotFound desc = could not find container \"5e3ea7b352e92ac01436df12b034e1b1c0e5c66742216f79fd703c305a9b2a3f\": container with ID starting with 5e3ea7b352e92ac01436df12b034e1b1c0e5c66742216f79fd703c305a9b2a3f not found: ID does not exist" Oct 02 11:13:19 crc kubenswrapper[4934]: I1002 11:13:19.113277 4934 scope.go:117] "RemoveContainer" containerID="dc55b5e2b06913bd776cc3cb499264b480ba13983dcfec8245b6b125888bda1e" Oct 02 11:13:19 crc kubenswrapper[4934]: E1002 11:13:19.113705 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc55b5e2b06913bd776cc3cb499264b480ba13983dcfec8245b6b125888bda1e\": container with ID starting with dc55b5e2b06913bd776cc3cb499264b480ba13983dcfec8245b6b125888bda1e not found: ID does not exist" containerID="dc55b5e2b06913bd776cc3cb499264b480ba13983dcfec8245b6b125888bda1e" Oct 02 11:13:19 crc kubenswrapper[4934]: I1002 11:13:19.113724 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc55b5e2b06913bd776cc3cb499264b480ba13983dcfec8245b6b125888bda1e"} err="failed to get container status \"dc55b5e2b06913bd776cc3cb499264b480ba13983dcfec8245b6b125888bda1e\": rpc error: code = NotFound desc = could not find container \"dc55b5e2b06913bd776cc3cb499264b480ba13983dcfec8245b6b125888bda1e\": container with ID starting with dc55b5e2b06913bd776cc3cb499264b480ba13983dcfec8245b6b125888bda1e not found: ID does not exist" Oct 02 11:13:19 crc kubenswrapper[4934]: I1002 11:13:19.113740 4934 scope.go:117] "RemoveContainer" containerID="adbf17d31317d364841f139cc80038389e30616329d51dabed6cf4d3cdebb4fd" Oct 02 11:13:19 crc kubenswrapper[4934]: E1002 11:13:19.114161 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"adbf17d31317d364841f139cc80038389e30616329d51dabed6cf4d3cdebb4fd\": container with ID starting with adbf17d31317d364841f139cc80038389e30616329d51dabed6cf4d3cdebb4fd not found: ID does not exist" containerID="adbf17d31317d364841f139cc80038389e30616329d51dabed6cf4d3cdebb4fd" Oct 02 11:13:19 crc kubenswrapper[4934]: I1002 11:13:19.114204 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adbf17d31317d364841f139cc80038389e30616329d51dabed6cf4d3cdebb4fd"} err="failed to get container status \"adbf17d31317d364841f139cc80038389e30616329d51dabed6cf4d3cdebb4fd\": rpc error: code = NotFound desc = could not find container \"adbf17d31317d364841f139cc80038389e30616329d51dabed6cf4d3cdebb4fd\": container with ID starting with adbf17d31317d364841f139cc80038389e30616329d51dabed6cf4d3cdebb4fd not found: ID does not exist" Oct 02 11:13:20 crc kubenswrapper[4934]: I1002 11:13:20.913950 4934 scope.go:117] "RemoveContainer" containerID="1d6ec196f0e4bccca35b31484223dd4730674c6d770e39fd96fd7a619d6ba562" Oct 02 11:13:20 crc kubenswrapper[4934]: E1002 11:13:20.914214 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:13:20 crc kubenswrapper[4934]: I1002 11:13:20.922086 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4365e38-5e5f-447f-a92c-d8b9870e6b31" path="/var/lib/kubelet/pods/b4365e38-5e5f-447f-a92c-d8b9870e6b31/volumes" Oct 02 11:13:29 crc kubenswrapper[4934]: E1002 11:13:29.691547 4934 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4c21786_bffe_40c4_9f2f_6f8046b5c1c0.slice/crio-conmon-a5f816b66778a9f91b5c93f72af6a24afd969975cad255559a107fa3238e2671.scope\": RecentStats: unable to find data in memory cache]" Oct 02 11:13:30 crc kubenswrapper[4934]: I1002 11:13:30.127363 4934 generic.go:334] "Generic (PLEG): container finished" podID="d4c21786-bffe-40c4-9f2f-6f8046b5c1c0" containerID="a5f816b66778a9f91b5c93f72af6a24afd969975cad255559a107fa3238e2671" exitCode=0 Oct 02 11:13:30 crc kubenswrapper[4934]: I1002 11:13:30.127417 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x6r6h/must-gather-9zg6r" event={"ID":"d4c21786-bffe-40c4-9f2f-6f8046b5c1c0","Type":"ContainerDied","Data":"a5f816b66778a9f91b5c93f72af6a24afd969975cad255559a107fa3238e2671"} Oct 02 11:13:30 crc kubenswrapper[4934]: I1002 11:13:30.127976 4934 scope.go:117] "RemoveContainer" containerID="a5f816b66778a9f91b5c93f72af6a24afd969975cad255559a107fa3238e2671" Oct 02 11:13:30 crc kubenswrapper[4934]: I1002 11:13:30.639542 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-x6r6h_must-gather-9zg6r_d4c21786-bffe-40c4-9f2f-6f8046b5c1c0/gather/0.log" Oct 02 11:13:33 crc kubenswrapper[4934]: I1002 11:13:33.913326 4934 scope.go:117] "RemoveContainer" containerID="1d6ec196f0e4bccca35b31484223dd4730674c6d770e39fd96fd7a619d6ba562" Oct 02 11:13:33 crc kubenswrapper[4934]: E1002 11:13:33.913745 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:13:37 crc kubenswrapper[4934]: I1002 11:13:37.764057 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-x6r6h/must-gather-9zg6r"] Oct 02 11:13:37 crc kubenswrapper[4934]: I1002 11:13:37.765005 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-x6r6h/must-gather-9zg6r" podUID="d4c21786-bffe-40c4-9f2f-6f8046b5c1c0" containerName="copy" containerID="cri-o://9e9176a751a90c013f39de73724866f30f1ec554ccfffdca838946ca95f745e9" gracePeriod=2 Oct 02 11:13:37 crc kubenswrapper[4934]: I1002 11:13:37.769173 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-x6r6h/must-gather-9zg6r"] Oct 02 11:13:38 crc kubenswrapper[4934]: I1002 11:13:38.180861 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-x6r6h_must-gather-9zg6r_d4c21786-bffe-40c4-9f2f-6f8046b5c1c0/copy/0.log" Oct 02 11:13:38 crc kubenswrapper[4934]: I1002 11:13:38.182493 4934 generic.go:334] "Generic (PLEG): container finished" podID="d4c21786-bffe-40c4-9f2f-6f8046b5c1c0" containerID="9e9176a751a90c013f39de73724866f30f1ec554ccfffdca838946ca95f745e9" exitCode=143 Oct 02 11:13:38 crc kubenswrapper[4934]: I1002 11:13:38.182709 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c68d07ab1948b92c2d54dc61c1bb9b613724b43311f08d3810199832afae749" Oct 02 11:13:38 crc kubenswrapper[4934]: I1002 11:13:38.207989 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-x6r6h_must-gather-9zg6r_d4c21786-bffe-40c4-9f2f-6f8046b5c1c0/copy/0.log" Oct 02 11:13:38 crc kubenswrapper[4934]: I1002 11:13:38.208383 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x6r6h/must-gather-9zg6r" Oct 02 11:13:38 crc kubenswrapper[4934]: I1002 11:13:38.275354 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d4c21786-bffe-40c4-9f2f-6f8046b5c1c0-must-gather-output\") pod \"d4c21786-bffe-40c4-9f2f-6f8046b5c1c0\" (UID: \"d4c21786-bffe-40c4-9f2f-6f8046b5c1c0\") " Oct 02 11:13:38 crc kubenswrapper[4934]: I1002 11:13:38.275408 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcnz7\" (UniqueName: \"kubernetes.io/projected/d4c21786-bffe-40c4-9f2f-6f8046b5c1c0-kube-api-access-rcnz7\") pod \"d4c21786-bffe-40c4-9f2f-6f8046b5c1c0\" (UID: \"d4c21786-bffe-40c4-9f2f-6f8046b5c1c0\") " Oct 02 11:13:38 crc kubenswrapper[4934]: I1002 11:13:38.285790 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4c21786-bffe-40c4-9f2f-6f8046b5c1c0-kube-api-access-rcnz7" (OuterVolumeSpecName: "kube-api-access-rcnz7") pod "d4c21786-bffe-40c4-9f2f-6f8046b5c1c0" (UID: "d4c21786-bffe-40c4-9f2f-6f8046b5c1c0"). InnerVolumeSpecName "kube-api-access-rcnz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:13:38 crc kubenswrapper[4934]: I1002 11:13:38.354180 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4c21786-bffe-40c4-9f2f-6f8046b5c1c0-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "d4c21786-bffe-40c4-9f2f-6f8046b5c1c0" (UID: "d4c21786-bffe-40c4-9f2f-6f8046b5c1c0"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:13:38 crc kubenswrapper[4934]: I1002 11:13:38.377051 4934 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d4c21786-bffe-40c4-9f2f-6f8046b5c1c0-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 02 11:13:38 crc kubenswrapper[4934]: I1002 11:13:38.377090 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcnz7\" (UniqueName: \"kubernetes.io/projected/d4c21786-bffe-40c4-9f2f-6f8046b5c1c0-kube-api-access-rcnz7\") on node \"crc\" DevicePath \"\"" Oct 02 11:13:38 crc kubenswrapper[4934]: I1002 11:13:38.931894 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4c21786-bffe-40c4-9f2f-6f8046b5c1c0" path="/var/lib/kubelet/pods/d4c21786-bffe-40c4-9f2f-6f8046b5c1c0/volumes" Oct 02 11:13:39 crc kubenswrapper[4934]: I1002 11:13:39.190154 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x6r6h/must-gather-9zg6r" Oct 02 11:13:48 crc kubenswrapper[4934]: I1002 11:13:48.917505 4934 scope.go:117] "RemoveContainer" containerID="1d6ec196f0e4bccca35b31484223dd4730674c6d770e39fd96fd7a619d6ba562" Oct 02 11:13:48 crc kubenswrapper[4934]: E1002 11:13:48.918823 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:14:03 crc kubenswrapper[4934]: I1002 11:14:03.913870 4934 scope.go:117] "RemoveContainer" containerID="1d6ec196f0e4bccca35b31484223dd4730674c6d770e39fd96fd7a619d6ba562" Oct 02 11:14:03 crc kubenswrapper[4934]: E1002 11:14:03.914851 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:14:17 crc kubenswrapper[4934]: I1002 11:14:17.913679 4934 scope.go:117] "RemoveContainer" containerID="1d6ec196f0e4bccca35b31484223dd4730674c6d770e39fd96fd7a619d6ba562" Oct 02 11:14:17 crc kubenswrapper[4934]: E1002 11:14:17.914384 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:14:31 crc kubenswrapper[4934]: I1002 11:14:31.913534 4934 scope.go:117] "RemoveContainer" containerID="1d6ec196f0e4bccca35b31484223dd4730674c6d770e39fd96fd7a619d6ba562" Oct 02 11:14:31 crc kubenswrapper[4934]: E1002 11:14:31.915676 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:14:37 crc kubenswrapper[4934]: I1002 11:14:37.238766 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tfhhq"] Oct 02 11:14:37 crc kubenswrapper[4934]: E1002 11:14:37.239434 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4c21786-bffe-40c4-9f2f-6f8046b5c1c0" containerName="copy" Oct 02 11:14:37 crc kubenswrapper[4934]: I1002 11:14:37.239448 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4c21786-bffe-40c4-9f2f-6f8046b5c1c0" containerName="copy" Oct 02 11:14:37 crc kubenswrapper[4934]: E1002 11:14:37.239466 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4c21786-bffe-40c4-9f2f-6f8046b5c1c0" containerName="gather" Oct 02 11:14:37 crc kubenswrapper[4934]: I1002 11:14:37.239473 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4c21786-bffe-40c4-9f2f-6f8046b5c1c0" containerName="gather" Oct 02 11:14:37 crc kubenswrapper[4934]: E1002 11:14:37.239484 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4365e38-5e5f-447f-a92c-d8b9870e6b31" containerName="registry-server" Oct 02 11:14:37 crc kubenswrapper[4934]: I1002 11:14:37.239492 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4365e38-5e5f-447f-a92c-d8b9870e6b31" containerName="registry-server" Oct 02 11:14:37 crc kubenswrapper[4934]: E1002 11:14:37.239518 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4365e38-5e5f-447f-a92c-d8b9870e6b31" containerName="extract-content" Oct 02 11:14:37 crc kubenswrapper[4934]: I1002 11:14:37.239526 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4365e38-5e5f-447f-a92c-d8b9870e6b31" containerName="extract-content" Oct 02 11:14:37 crc kubenswrapper[4934]: E1002 11:14:37.239544 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4365e38-5e5f-447f-a92c-d8b9870e6b31" containerName="extract-utilities" Oct 02 11:14:37 crc kubenswrapper[4934]: I1002 11:14:37.239551 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4365e38-5e5f-447f-a92c-d8b9870e6b31" containerName="extract-utilities" Oct 02 11:14:37 crc kubenswrapper[4934]: I1002 11:14:37.239745 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4c21786-bffe-40c4-9f2f-6f8046b5c1c0" containerName="copy" Oct 02 11:14:37 crc kubenswrapper[4934]: I1002 11:14:37.239765 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4c21786-bffe-40c4-9f2f-6f8046b5c1c0" containerName="gather" Oct 02 11:14:37 crc kubenswrapper[4934]: I1002 11:14:37.239783 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4365e38-5e5f-447f-a92c-d8b9870e6b31" containerName="registry-server" Oct 02 11:14:37 crc kubenswrapper[4934]: I1002 11:14:37.241027 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tfhhq" Oct 02 11:14:37 crc kubenswrapper[4934]: I1002 11:14:37.257991 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tfhhq"] Oct 02 11:14:37 crc kubenswrapper[4934]: I1002 11:14:37.409968 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20023882-f372-4d02-9719-e8775649ad66-catalog-content\") pod \"community-operators-tfhhq\" (UID: \"20023882-f372-4d02-9719-e8775649ad66\") " pod="openshift-marketplace/community-operators-tfhhq" Oct 02 11:14:37 crc kubenswrapper[4934]: I1002 11:14:37.410280 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7xwd\" (UniqueName: \"kubernetes.io/projected/20023882-f372-4d02-9719-e8775649ad66-kube-api-access-g7xwd\") pod \"community-operators-tfhhq\" (UID: \"20023882-f372-4d02-9719-e8775649ad66\") " pod="openshift-marketplace/community-operators-tfhhq" Oct 02 11:14:37 crc kubenswrapper[4934]: I1002 11:14:37.410421 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20023882-f372-4d02-9719-e8775649ad66-utilities\") pod \"community-operators-tfhhq\" (UID: \"20023882-f372-4d02-9719-e8775649ad66\") " pod="openshift-marketplace/community-operators-tfhhq" Oct 02 11:14:37 crc kubenswrapper[4934]: I1002 11:14:37.511406 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20023882-f372-4d02-9719-e8775649ad66-catalog-content\") pod \"community-operators-tfhhq\" (UID: \"20023882-f372-4d02-9719-e8775649ad66\") " pod="openshift-marketplace/community-operators-tfhhq" Oct 02 11:14:37 crc kubenswrapper[4934]: I1002 11:14:37.511745 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7xwd\" (UniqueName: \"kubernetes.io/projected/20023882-f372-4d02-9719-e8775649ad66-kube-api-access-g7xwd\") pod \"community-operators-tfhhq\" (UID: \"20023882-f372-4d02-9719-e8775649ad66\") " pod="openshift-marketplace/community-operators-tfhhq" Oct 02 11:14:37 crc kubenswrapper[4934]: I1002 11:14:37.511840 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20023882-f372-4d02-9719-e8775649ad66-utilities\") pod \"community-operators-tfhhq\" (UID: \"20023882-f372-4d02-9719-e8775649ad66\") " pod="openshift-marketplace/community-operators-tfhhq" Oct 02 11:14:37 crc kubenswrapper[4934]: I1002 11:14:37.511889 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20023882-f372-4d02-9719-e8775649ad66-catalog-content\") pod \"community-operators-tfhhq\" (UID: \"20023882-f372-4d02-9719-e8775649ad66\") " pod="openshift-marketplace/community-operators-tfhhq" Oct 02 11:14:37 crc kubenswrapper[4934]: I1002 11:14:37.512423 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20023882-f372-4d02-9719-e8775649ad66-utilities\") pod \"community-operators-tfhhq\" (UID: \"20023882-f372-4d02-9719-e8775649ad66\") " pod="openshift-marketplace/community-operators-tfhhq" Oct 02 11:14:37 crc kubenswrapper[4934]: I1002 11:14:37.530786 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7xwd\" (UniqueName: \"kubernetes.io/projected/20023882-f372-4d02-9719-e8775649ad66-kube-api-access-g7xwd\") pod \"community-operators-tfhhq\" (UID: \"20023882-f372-4d02-9719-e8775649ad66\") " pod="openshift-marketplace/community-operators-tfhhq" Oct 02 11:14:37 crc kubenswrapper[4934]: I1002 11:14:37.558485 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tfhhq" Oct 02 11:14:38 crc kubenswrapper[4934]: I1002 11:14:38.070340 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tfhhq"] Oct 02 11:14:38 crc kubenswrapper[4934]: W1002 11:14:38.077752 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20023882_f372_4d02_9719_e8775649ad66.slice/crio-750e1f820e2ae2198330a5f40486fb40d8b2629a9ce1bb95ad1a5d5163dd5703 WatchSource:0}: Error finding container 750e1f820e2ae2198330a5f40486fb40d8b2629a9ce1bb95ad1a5d5163dd5703: Status 404 returned error can't find the container with id 750e1f820e2ae2198330a5f40486fb40d8b2629a9ce1bb95ad1a5d5163dd5703 Oct 02 11:14:38 crc kubenswrapper[4934]: I1002 11:14:38.640667 4934 generic.go:334] "Generic (PLEG): container finished" podID="20023882-f372-4d02-9719-e8775649ad66" containerID="a2cbd6800163ba87b02b19a3c86110e240eeb5d1f2781bb80c4f73a50ea2feab" exitCode=0 Oct 02 11:14:38 crc kubenswrapper[4934]: I1002 11:14:38.640735 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tfhhq" event={"ID":"20023882-f372-4d02-9719-e8775649ad66","Type":"ContainerDied","Data":"a2cbd6800163ba87b02b19a3c86110e240eeb5d1f2781bb80c4f73a50ea2feab"} Oct 02 11:14:38 crc kubenswrapper[4934]: I1002 11:14:38.640921 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tfhhq" event={"ID":"20023882-f372-4d02-9719-e8775649ad66","Type":"ContainerStarted","Data":"750e1f820e2ae2198330a5f40486fb40d8b2629a9ce1bb95ad1a5d5163dd5703"} Oct 02 11:14:38 crc kubenswrapper[4934]: I1002 11:14:38.643944 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 11:14:39 crc kubenswrapper[4934]: I1002 11:14:39.654359 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tfhhq" event={"ID":"20023882-f372-4d02-9719-e8775649ad66","Type":"ContainerStarted","Data":"5a81d89141e594fc9016dbe26d6dcea790618752daf7acbc0754f53cd7dff401"} Oct 02 11:14:40 crc kubenswrapper[4934]: I1002 11:14:40.664390 4934 generic.go:334] "Generic (PLEG): container finished" podID="20023882-f372-4d02-9719-e8775649ad66" containerID="5a81d89141e594fc9016dbe26d6dcea790618752daf7acbc0754f53cd7dff401" exitCode=0 Oct 02 11:14:40 crc kubenswrapper[4934]: I1002 11:14:40.664440 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tfhhq" event={"ID":"20023882-f372-4d02-9719-e8775649ad66","Type":"ContainerDied","Data":"5a81d89141e594fc9016dbe26d6dcea790618752daf7acbc0754f53cd7dff401"} Oct 02 11:14:41 crc kubenswrapper[4934]: I1002 11:14:41.671989 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tfhhq" event={"ID":"20023882-f372-4d02-9719-e8775649ad66","Type":"ContainerStarted","Data":"fd28232676f204ff450c3a7087b48df079149fcab755420c24f12b547635fa4b"} Oct 02 11:14:41 crc kubenswrapper[4934]: I1002 11:14:41.688523 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tfhhq" podStartSLOduration=2.077524295 podStartE2EDuration="4.688503607s" podCreationTimestamp="2025-10-02 11:14:37 +0000 UTC" firstStartedPulling="2025-10-02 11:14:38.643723939 +0000 UTC m=+5150.396365461" lastFinishedPulling="2025-10-02 11:14:41.254703251 +0000 UTC m=+5153.007344773" observedRunningTime="2025-10-02 11:14:41.686282027 +0000 UTC m=+5153.438923569" watchObservedRunningTime="2025-10-02 11:14:41.688503607 +0000 UTC m=+5153.441145129" Oct 02 11:14:44 crc kubenswrapper[4934]: I1002 11:14:44.913803 4934 scope.go:117] "RemoveContainer" containerID="1d6ec196f0e4bccca35b31484223dd4730674c6d770e39fd96fd7a619d6ba562" Oct 02 11:14:44 crc kubenswrapper[4934]: E1002 11:14:44.914517 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:14:47 crc kubenswrapper[4934]: I1002 11:14:47.559286 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tfhhq" Oct 02 11:14:47 crc kubenswrapper[4934]: I1002 11:14:47.559355 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tfhhq" Oct 02 11:14:47 crc kubenswrapper[4934]: I1002 11:14:47.636892 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tfhhq" Oct 02 11:14:47 crc kubenswrapper[4934]: I1002 11:14:47.765454 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tfhhq" Oct 02 11:14:47 crc kubenswrapper[4934]: I1002 11:14:47.869749 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tfhhq"] Oct 02 11:14:49 crc kubenswrapper[4934]: I1002 11:14:49.730930 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tfhhq" podUID="20023882-f372-4d02-9719-e8775649ad66" containerName="registry-server" containerID="cri-o://fd28232676f204ff450c3a7087b48df079149fcab755420c24f12b547635fa4b" gracePeriod=2 Oct 02 11:14:50 crc kubenswrapper[4934]: I1002 11:14:50.159095 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tfhhq" Oct 02 11:14:50 crc kubenswrapper[4934]: I1002 11:14:50.288284 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20023882-f372-4d02-9719-e8775649ad66-utilities\") pod \"20023882-f372-4d02-9719-e8775649ad66\" (UID: \"20023882-f372-4d02-9719-e8775649ad66\") " Oct 02 11:14:50 crc kubenswrapper[4934]: I1002 11:14:50.288615 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7xwd\" (UniqueName: \"kubernetes.io/projected/20023882-f372-4d02-9719-e8775649ad66-kube-api-access-g7xwd\") pod \"20023882-f372-4d02-9719-e8775649ad66\" (UID: \"20023882-f372-4d02-9719-e8775649ad66\") " Oct 02 11:14:50 crc kubenswrapper[4934]: I1002 11:14:50.288709 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20023882-f372-4d02-9719-e8775649ad66-catalog-content\") pod \"20023882-f372-4d02-9719-e8775649ad66\" (UID: \"20023882-f372-4d02-9719-e8775649ad66\") " Oct 02 11:14:50 crc kubenswrapper[4934]: I1002 11:14:50.289344 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20023882-f372-4d02-9719-e8775649ad66-utilities" (OuterVolumeSpecName: "utilities") pod "20023882-f372-4d02-9719-e8775649ad66" (UID: "20023882-f372-4d02-9719-e8775649ad66"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:14:50 crc kubenswrapper[4934]: I1002 11:14:50.298950 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20023882-f372-4d02-9719-e8775649ad66-kube-api-access-g7xwd" (OuterVolumeSpecName: "kube-api-access-g7xwd") pod "20023882-f372-4d02-9719-e8775649ad66" (UID: "20023882-f372-4d02-9719-e8775649ad66"). InnerVolumeSpecName "kube-api-access-g7xwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:14:50 crc kubenswrapper[4934]: I1002 11:14:50.398117 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7xwd\" (UniqueName: \"kubernetes.io/projected/20023882-f372-4d02-9719-e8775649ad66-kube-api-access-g7xwd\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:50 crc kubenswrapper[4934]: I1002 11:14:50.398182 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20023882-f372-4d02-9719-e8775649ad66-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:50 crc kubenswrapper[4934]: I1002 11:14:50.740640 4934 generic.go:334] "Generic (PLEG): container finished" podID="20023882-f372-4d02-9719-e8775649ad66" containerID="fd28232676f204ff450c3a7087b48df079149fcab755420c24f12b547635fa4b" exitCode=0 Oct 02 11:14:50 crc kubenswrapper[4934]: I1002 11:14:50.740704 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tfhhq" event={"ID":"20023882-f372-4d02-9719-e8775649ad66","Type":"ContainerDied","Data":"fd28232676f204ff450c3a7087b48df079149fcab755420c24f12b547635fa4b"} Oct 02 11:14:50 crc kubenswrapper[4934]: I1002 11:14:50.740726 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tfhhq" Oct 02 11:14:50 crc kubenswrapper[4934]: I1002 11:14:50.741467 4934 scope.go:117] "RemoveContainer" containerID="fd28232676f204ff450c3a7087b48df079149fcab755420c24f12b547635fa4b" Oct 02 11:14:50 crc kubenswrapper[4934]: I1002 11:14:50.741381 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tfhhq" event={"ID":"20023882-f372-4d02-9719-e8775649ad66","Type":"ContainerDied","Data":"750e1f820e2ae2198330a5f40486fb40d8b2629a9ce1bb95ad1a5d5163dd5703"} Oct 02 11:14:50 crc kubenswrapper[4934]: I1002 11:14:50.766733 4934 scope.go:117] "RemoveContainer" containerID="5a81d89141e594fc9016dbe26d6dcea790618752daf7acbc0754f53cd7dff401" Oct 02 11:14:50 crc kubenswrapper[4934]: I1002 11:14:50.788864 4934 scope.go:117] "RemoveContainer" containerID="a2cbd6800163ba87b02b19a3c86110e240eeb5d1f2781bb80c4f73a50ea2feab" Oct 02 11:14:50 crc kubenswrapper[4934]: I1002 11:14:50.813458 4934 scope.go:117] "RemoveContainer" containerID="fd28232676f204ff450c3a7087b48df079149fcab755420c24f12b547635fa4b" Oct 02 11:14:50 crc kubenswrapper[4934]: E1002 11:14:50.814123 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd28232676f204ff450c3a7087b48df079149fcab755420c24f12b547635fa4b\": container with ID starting with fd28232676f204ff450c3a7087b48df079149fcab755420c24f12b547635fa4b not found: ID does not exist" containerID="fd28232676f204ff450c3a7087b48df079149fcab755420c24f12b547635fa4b" Oct 02 11:14:50 crc kubenswrapper[4934]: I1002 11:14:50.814178 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd28232676f204ff450c3a7087b48df079149fcab755420c24f12b547635fa4b"} err="failed to get container status \"fd28232676f204ff450c3a7087b48df079149fcab755420c24f12b547635fa4b\": rpc error: code = NotFound desc = could not find container \"fd28232676f204ff450c3a7087b48df079149fcab755420c24f12b547635fa4b\": container with ID starting with fd28232676f204ff450c3a7087b48df079149fcab755420c24f12b547635fa4b not found: ID does not exist" Oct 02 11:14:50 crc kubenswrapper[4934]: I1002 11:14:50.814213 4934 scope.go:117] "RemoveContainer" containerID="5a81d89141e594fc9016dbe26d6dcea790618752daf7acbc0754f53cd7dff401" Oct 02 11:14:50 crc kubenswrapper[4934]: E1002 11:14:50.814626 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a81d89141e594fc9016dbe26d6dcea790618752daf7acbc0754f53cd7dff401\": container with ID starting with 5a81d89141e594fc9016dbe26d6dcea790618752daf7acbc0754f53cd7dff401 not found: ID does not exist" containerID="5a81d89141e594fc9016dbe26d6dcea790618752daf7acbc0754f53cd7dff401" Oct 02 11:14:50 crc kubenswrapper[4934]: I1002 11:14:50.814657 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a81d89141e594fc9016dbe26d6dcea790618752daf7acbc0754f53cd7dff401"} err="failed to get container status \"5a81d89141e594fc9016dbe26d6dcea790618752daf7acbc0754f53cd7dff401\": rpc error: code = NotFound desc = could not find container \"5a81d89141e594fc9016dbe26d6dcea790618752daf7acbc0754f53cd7dff401\": container with ID starting with 5a81d89141e594fc9016dbe26d6dcea790618752daf7acbc0754f53cd7dff401 not found: ID does not exist" Oct 02 11:14:50 crc kubenswrapper[4934]: I1002 11:14:50.814675 4934 scope.go:117] "RemoveContainer" containerID="a2cbd6800163ba87b02b19a3c86110e240eeb5d1f2781bb80c4f73a50ea2feab" Oct 02 11:14:50 crc kubenswrapper[4934]: E1002 11:14:50.814971 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2cbd6800163ba87b02b19a3c86110e240eeb5d1f2781bb80c4f73a50ea2feab\": container with ID starting with a2cbd6800163ba87b02b19a3c86110e240eeb5d1f2781bb80c4f73a50ea2feab not found: ID does not exist" containerID="a2cbd6800163ba87b02b19a3c86110e240eeb5d1f2781bb80c4f73a50ea2feab" Oct 02 11:14:50 crc kubenswrapper[4934]: I1002 11:14:50.815003 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2cbd6800163ba87b02b19a3c86110e240eeb5d1f2781bb80c4f73a50ea2feab"} err="failed to get container status \"a2cbd6800163ba87b02b19a3c86110e240eeb5d1f2781bb80c4f73a50ea2feab\": rpc error: code = NotFound desc = could not find container \"a2cbd6800163ba87b02b19a3c86110e240eeb5d1f2781bb80c4f73a50ea2feab\": container with ID starting with a2cbd6800163ba87b02b19a3c86110e240eeb5d1f2781bb80c4f73a50ea2feab not found: ID does not exist" Oct 02 11:14:50 crc kubenswrapper[4934]: I1002 11:14:50.976950 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20023882-f372-4d02-9719-e8775649ad66-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "20023882-f372-4d02-9719-e8775649ad66" (UID: "20023882-f372-4d02-9719-e8775649ad66"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 11:14:51 crc kubenswrapper[4934]: I1002 11:14:51.012631 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20023882-f372-4d02-9719-e8775649ad66-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 11:14:51 crc kubenswrapper[4934]: I1002 11:14:51.073304 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tfhhq"] Oct 02 11:14:51 crc kubenswrapper[4934]: I1002 11:14:51.080775 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tfhhq"] Oct 02 11:14:52 crc kubenswrapper[4934]: I1002 11:14:52.923337 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20023882-f372-4d02-9719-e8775649ad66" path="/var/lib/kubelet/pods/20023882-f372-4d02-9719-e8775649ad66/volumes" Oct 02 11:14:56 crc kubenswrapper[4934]: I1002 11:14:56.912882 4934 scope.go:117] "RemoveContainer" containerID="1d6ec196f0e4bccca35b31484223dd4730674c6d770e39fd96fd7a619d6ba562" Oct 02 11:14:56 crc kubenswrapper[4934]: E1002 11:14:56.913403 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:15:00 crc kubenswrapper[4934]: I1002 11:15:00.147163 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323395-2xksr"] Oct 02 11:15:00 crc kubenswrapper[4934]: E1002 11:15:00.147616 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20023882-f372-4d02-9719-e8775649ad66" containerName="extract-utilities" Oct 02 11:15:00 crc kubenswrapper[4934]: I1002 11:15:00.147642 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="20023882-f372-4d02-9719-e8775649ad66" containerName="extract-utilities" Oct 02 11:15:00 crc kubenswrapper[4934]: E1002 11:15:00.147666 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20023882-f372-4d02-9719-e8775649ad66" containerName="registry-server" Oct 02 11:15:00 crc kubenswrapper[4934]: I1002 11:15:00.147678 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="20023882-f372-4d02-9719-e8775649ad66" containerName="registry-server" Oct 02 11:15:00 crc kubenswrapper[4934]: E1002 11:15:00.147699 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20023882-f372-4d02-9719-e8775649ad66" containerName="extract-content" Oct 02 11:15:00 crc kubenswrapper[4934]: I1002 11:15:00.147713 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="20023882-f372-4d02-9719-e8775649ad66" containerName="extract-content" Oct 02 11:15:00 crc kubenswrapper[4934]: I1002 11:15:00.147994 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="20023882-f372-4d02-9719-e8775649ad66" containerName="registry-server" Oct 02 11:15:00 crc kubenswrapper[4934]: I1002 11:15:00.148731 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-2xksr" Oct 02 11:15:00 crc kubenswrapper[4934]: I1002 11:15:00.150725 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 11:15:00 crc kubenswrapper[4934]: I1002 11:15:00.151045 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 11:15:00 crc kubenswrapper[4934]: I1002 11:15:00.154420 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323395-2xksr"] Oct 02 11:15:00 crc kubenswrapper[4934]: I1002 11:15:00.237794 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e790249c-3b9d-420e-9dad-847f4bcaefca-config-volume\") pod \"collect-profiles-29323395-2xksr\" (UID: \"e790249c-3b9d-420e-9dad-847f4bcaefca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-2xksr" Oct 02 11:15:00 crc kubenswrapper[4934]: I1002 11:15:00.238813 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26g8f\" (UniqueName: \"kubernetes.io/projected/e790249c-3b9d-420e-9dad-847f4bcaefca-kube-api-access-26g8f\") pod \"collect-profiles-29323395-2xksr\" (UID: \"e790249c-3b9d-420e-9dad-847f4bcaefca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-2xksr" Oct 02 11:15:00 crc kubenswrapper[4934]: I1002 11:15:00.238893 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e790249c-3b9d-420e-9dad-847f4bcaefca-secret-volume\") pod \"collect-profiles-29323395-2xksr\" (UID: \"e790249c-3b9d-420e-9dad-847f4bcaefca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-2xksr" Oct 02 11:15:00 crc kubenswrapper[4934]: I1002 11:15:00.339888 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e790249c-3b9d-420e-9dad-847f4bcaefca-config-volume\") pod \"collect-profiles-29323395-2xksr\" (UID: \"e790249c-3b9d-420e-9dad-847f4bcaefca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-2xksr" Oct 02 11:15:00 crc kubenswrapper[4934]: I1002 11:15:00.339958 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26g8f\" (UniqueName: \"kubernetes.io/projected/e790249c-3b9d-420e-9dad-847f4bcaefca-kube-api-access-26g8f\") pod \"collect-profiles-29323395-2xksr\" (UID: \"e790249c-3b9d-420e-9dad-847f4bcaefca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-2xksr" Oct 02 11:15:00 crc kubenswrapper[4934]: I1002 11:15:00.339985 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e790249c-3b9d-420e-9dad-847f4bcaefca-secret-volume\") pod \"collect-profiles-29323395-2xksr\" (UID: \"e790249c-3b9d-420e-9dad-847f4bcaefca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-2xksr" Oct 02 11:15:00 crc kubenswrapper[4934]: I1002 11:15:00.341066 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e790249c-3b9d-420e-9dad-847f4bcaefca-config-volume\") pod \"collect-profiles-29323395-2xksr\" (UID: \"e790249c-3b9d-420e-9dad-847f4bcaefca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-2xksr" Oct 02 11:15:00 crc kubenswrapper[4934]: I1002 11:15:00.345333 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e790249c-3b9d-420e-9dad-847f4bcaefca-secret-volume\") pod \"collect-profiles-29323395-2xksr\" (UID: \"e790249c-3b9d-420e-9dad-847f4bcaefca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-2xksr" Oct 02 11:15:00 crc kubenswrapper[4934]: I1002 11:15:00.355267 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26g8f\" (UniqueName: \"kubernetes.io/projected/e790249c-3b9d-420e-9dad-847f4bcaefca-kube-api-access-26g8f\") pod \"collect-profiles-29323395-2xksr\" (UID: \"e790249c-3b9d-420e-9dad-847f4bcaefca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-2xksr" Oct 02 11:15:00 crc kubenswrapper[4934]: I1002 11:15:00.470526 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-2xksr" Oct 02 11:15:00 crc kubenswrapper[4934]: I1002 11:15:00.867394 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323395-2xksr"] Oct 02 11:15:01 crc kubenswrapper[4934]: I1002 11:15:01.813179 4934 generic.go:334] "Generic (PLEG): container finished" podID="e790249c-3b9d-420e-9dad-847f4bcaefca" containerID="e3b946fe8bce65f80453492a6460b56055c1ef71f2e4fe6debb757ed14e5f97a" exitCode=0 Oct 02 11:15:01 crc kubenswrapper[4934]: I1002 11:15:01.813226 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-2xksr" event={"ID":"e790249c-3b9d-420e-9dad-847f4bcaefca","Type":"ContainerDied","Data":"e3b946fe8bce65f80453492a6460b56055c1ef71f2e4fe6debb757ed14e5f97a"} Oct 02 11:15:01 crc kubenswrapper[4934]: I1002 11:15:01.813257 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-2xksr" event={"ID":"e790249c-3b9d-420e-9dad-847f4bcaefca","Type":"ContainerStarted","Data":"bb673d77b5f29757cca7380fdacb66226201a2525c6195f67d6b8c3b9d36943a"} Oct 02 11:15:03 crc kubenswrapper[4934]: I1002 11:15:03.110975 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-2xksr" Oct 02 11:15:03 crc kubenswrapper[4934]: I1002 11:15:03.273674 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e790249c-3b9d-420e-9dad-847f4bcaefca-config-volume\") pod \"e790249c-3b9d-420e-9dad-847f4bcaefca\" (UID: \"e790249c-3b9d-420e-9dad-847f4bcaefca\") " Oct 02 11:15:03 crc kubenswrapper[4934]: I1002 11:15:03.274167 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26g8f\" (UniqueName: \"kubernetes.io/projected/e790249c-3b9d-420e-9dad-847f4bcaefca-kube-api-access-26g8f\") pod \"e790249c-3b9d-420e-9dad-847f4bcaefca\" (UID: \"e790249c-3b9d-420e-9dad-847f4bcaefca\") " Oct 02 11:15:03 crc kubenswrapper[4934]: I1002 11:15:03.274234 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e790249c-3b9d-420e-9dad-847f4bcaefca-secret-volume\") pod \"e790249c-3b9d-420e-9dad-847f4bcaefca\" (UID: \"e790249c-3b9d-420e-9dad-847f4bcaefca\") " Oct 02 11:15:03 crc kubenswrapper[4934]: I1002 11:15:03.275307 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e790249c-3b9d-420e-9dad-847f4bcaefca-config-volume" (OuterVolumeSpecName: "config-volume") pod "e790249c-3b9d-420e-9dad-847f4bcaefca" (UID: "e790249c-3b9d-420e-9dad-847f4bcaefca"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 11:15:03 crc kubenswrapper[4934]: I1002 11:15:03.284032 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e790249c-3b9d-420e-9dad-847f4bcaefca-kube-api-access-26g8f" (OuterVolumeSpecName: "kube-api-access-26g8f") pod "e790249c-3b9d-420e-9dad-847f4bcaefca" (UID: "e790249c-3b9d-420e-9dad-847f4bcaefca"). InnerVolumeSpecName "kube-api-access-26g8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 11:15:03 crc kubenswrapper[4934]: I1002 11:15:03.284071 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e790249c-3b9d-420e-9dad-847f4bcaefca-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e790249c-3b9d-420e-9dad-847f4bcaefca" (UID: "e790249c-3b9d-420e-9dad-847f4bcaefca"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 11:15:03 crc kubenswrapper[4934]: I1002 11:15:03.375630 4934 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e790249c-3b9d-420e-9dad-847f4bcaefca-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:03 crc kubenswrapper[4934]: I1002 11:15:03.375679 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26g8f\" (UniqueName: \"kubernetes.io/projected/e790249c-3b9d-420e-9dad-847f4bcaefca-kube-api-access-26g8f\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:03 crc kubenswrapper[4934]: I1002 11:15:03.375696 4934 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e790249c-3b9d-420e-9dad-847f4bcaefca-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 11:15:03 crc kubenswrapper[4934]: I1002 11:15:03.828649 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-2xksr" event={"ID":"e790249c-3b9d-420e-9dad-847f4bcaefca","Type":"ContainerDied","Data":"bb673d77b5f29757cca7380fdacb66226201a2525c6195f67d6b8c3b9d36943a"} Oct 02 11:15:03 crc kubenswrapper[4934]: I1002 11:15:03.828685 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb673d77b5f29757cca7380fdacb66226201a2525c6195f67d6b8c3b9d36943a" Oct 02 11:15:03 crc kubenswrapper[4934]: I1002 11:15:03.828714 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323395-2xksr" Oct 02 11:15:04 crc kubenswrapper[4934]: I1002 11:15:04.196597 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323350-wh248"] Oct 02 11:15:04 crc kubenswrapper[4934]: I1002 11:15:04.202481 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323350-wh248"] Oct 02 11:15:04 crc kubenswrapper[4934]: I1002 11:15:04.924087 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a87f0aa-7ba8-44dd-b72d-e0c235640e0d" path="/var/lib/kubelet/pods/4a87f0aa-7ba8-44dd-b72d-e0c235640e0d/volumes" Oct 02 11:15:10 crc kubenswrapper[4934]: I1002 11:15:10.913195 4934 scope.go:117] "RemoveContainer" containerID="1d6ec196f0e4bccca35b31484223dd4730674c6d770e39fd96fd7a619d6ba562" Oct 02 11:15:10 crc kubenswrapper[4934]: E1002 11:15:10.914153 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" Oct 02 11:15:14 crc kubenswrapper[4934]: I1002 11:15:14.628263 4934 scope.go:117] "RemoveContainer" containerID="53ef02571bceaeaf727a911421e4f6ce17d8e14cc898d3fad39ce0b0dc0ff551" Oct 02 11:15:22 crc kubenswrapper[4934]: I1002 11:15:22.913108 4934 scope.go:117] "RemoveContainer" containerID="1d6ec196f0e4bccca35b31484223dd4730674c6d770e39fd96fd7a619d6ba562" Oct 02 11:15:22 crc kubenswrapper[4934]: E1002 11:15:22.913836 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-djh5z_openshift-machine-config-operator(71db06ef-05b0-4f58-b251-b27117a8500a)\"" pod="openshift-machine-config-operator/machine-config-daemon-djh5z" podUID="71db06ef-05b0-4f58-b251-b27117a8500a" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515067457330024457 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015067457331017375 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015067444601016514 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015067444602015465 5ustar corecore